hexsha
string
size
int64
ext
string
lang
string
max_stars_repo_path
string
max_stars_repo_name
string
max_stars_repo_head_hexsha
string
max_stars_repo_licenses
list
max_stars_count
int64
max_stars_repo_stars_event_min_datetime
string
max_stars_repo_stars_event_max_datetime
string
max_issues_repo_path
string
max_issues_repo_name
string
max_issues_repo_head_hexsha
string
max_issues_repo_licenses
list
max_issues_count
int64
max_issues_repo_issues_event_min_datetime
string
max_issues_repo_issues_event_max_datetime
string
max_forks_repo_path
string
max_forks_repo_name
string
max_forks_repo_head_hexsha
string
max_forks_repo_licenses
list
max_forks_count
int64
max_forks_repo_forks_event_min_datetime
string
max_forks_repo_forks_event_max_datetime
string
content
string
avg_line_length
float64
max_line_length
int64
alphanum_fraction
float64
qsc_code_num_words_quality_signal
int64
qsc_code_num_chars_quality_signal
float64
qsc_code_mean_word_length_quality_signal
float64
qsc_code_frac_words_unique_quality_signal
float64
qsc_code_frac_chars_top_2grams_quality_signal
float64
qsc_code_frac_chars_top_3grams_quality_signal
float64
qsc_code_frac_chars_top_4grams_quality_signal
float64
qsc_code_frac_chars_dupe_5grams_quality_signal
float64
qsc_code_frac_chars_dupe_6grams_quality_signal
float64
qsc_code_frac_chars_dupe_7grams_quality_signal
float64
qsc_code_frac_chars_dupe_8grams_quality_signal
float64
qsc_code_frac_chars_dupe_9grams_quality_signal
float64
qsc_code_frac_chars_dupe_10grams_quality_signal
float64
qsc_code_frac_chars_replacement_symbols_quality_signal
float64
qsc_code_frac_chars_digital_quality_signal
float64
qsc_code_frac_chars_whitespace_quality_signal
float64
qsc_code_size_file_byte_quality_signal
float64
qsc_code_num_lines_quality_signal
float64
qsc_code_num_chars_line_max_quality_signal
float64
qsc_code_num_chars_line_mean_quality_signal
float64
qsc_code_frac_chars_alphabet_quality_signal
float64
qsc_code_frac_chars_comments_quality_signal
float64
qsc_code_cate_xml_start_quality_signal
float64
qsc_code_frac_lines_dupe_lines_quality_signal
float64
qsc_code_cate_autogen_quality_signal
float64
qsc_code_frac_lines_long_string_quality_signal
float64
qsc_code_frac_chars_string_length_quality_signal
float64
qsc_code_frac_chars_long_word_length_quality_signal
float64
qsc_code_frac_lines_string_concat_quality_signal
float64
qsc_code_cate_encoded_data_quality_signal
float64
qsc_code_frac_chars_hex_words_quality_signal
float64
qsc_code_frac_lines_prompt_comments_quality_signal
float64
qsc_code_frac_lines_assert_quality_signal
float64
qsc_codepython_cate_ast_quality_signal
float64
qsc_codepython_frac_lines_func_ratio_quality_signal
float64
qsc_codepython_cate_var_zero_quality_signal
bool
qsc_codepython_frac_lines_pass_quality_signal
float64
qsc_codepython_frac_lines_import_quality_signal
float64
qsc_codepython_frac_lines_simplefunc_quality_signal
float64
qsc_codepython_score_lines_no_logic_quality_signal
float64
qsc_codepython_frac_lines_print_quality_signal
float64
qsc_code_num_words
int64
qsc_code_num_chars
int64
qsc_code_mean_word_length
int64
qsc_code_frac_words_unique
null
qsc_code_frac_chars_top_2grams
int64
qsc_code_frac_chars_top_3grams
int64
qsc_code_frac_chars_top_4grams
int64
qsc_code_frac_chars_dupe_5grams
int64
qsc_code_frac_chars_dupe_6grams
int64
qsc_code_frac_chars_dupe_7grams
int64
qsc_code_frac_chars_dupe_8grams
int64
qsc_code_frac_chars_dupe_9grams
int64
qsc_code_frac_chars_dupe_10grams
int64
qsc_code_frac_chars_replacement_symbols
int64
qsc_code_frac_chars_digital
int64
qsc_code_frac_chars_whitespace
int64
qsc_code_size_file_byte
int64
qsc_code_num_lines
int64
qsc_code_num_chars_line_max
int64
qsc_code_num_chars_line_mean
int64
qsc_code_frac_chars_alphabet
int64
qsc_code_frac_chars_comments
int64
qsc_code_cate_xml_start
int64
qsc_code_frac_lines_dupe_lines
int64
qsc_code_cate_autogen
int64
qsc_code_frac_lines_long_string
int64
qsc_code_frac_chars_string_length
int64
qsc_code_frac_chars_long_word_length
int64
qsc_code_frac_lines_string_concat
null
qsc_code_cate_encoded_data
int64
qsc_code_frac_chars_hex_words
int64
qsc_code_frac_lines_prompt_comments
int64
qsc_code_frac_lines_assert
int64
qsc_codepython_cate_ast
int64
qsc_codepython_frac_lines_func_ratio
int64
qsc_codepython_cate_var_zero
int64
qsc_codepython_frac_lines_pass
int64
qsc_codepython_frac_lines_import
int64
qsc_codepython_frac_lines_simplefunc
int64
qsc_codepython_score_lines_no_logic
int64
qsc_codepython_frac_lines_print
int64
effective
string
hits
int64
a8f8c437318e16ccc0b9047af193be205971e26a
23,605
py
Python
aitoolbox/torchtrain/train_loop/train_loop_tracking.py
mv1388/AIToolbox
c64ac4810a02d230ce471d86b758e82ea232a7e7
[ "MIT" ]
null
null
null
aitoolbox/torchtrain/train_loop/train_loop_tracking.py
mv1388/AIToolbox
c64ac4810a02d230ce471d86b758e82ea232a7e7
[ "MIT" ]
1
2019-07-13T21:12:02.000Z
2019-07-13T21:12:02.000Z
aitoolbox/torchtrain/train_loop/train_loop_tracking.py
mv1388/AIToolbox
c64ac4810a02d230ce471d86b758e82ea232a7e7
[ "MIT" ]
null
null
null
import os import inspect from aitoolbox.torchtrain.train_loop.train_loop import TrainLoop from aitoolbox.experiment.result_package.abstract_result_packages import AbstractResultPackage from aitoolbox.torchtrain.callbacks.model_save import ModelCheckpoint, ModelIterationCheckpoint, ModelTrainEndSave from aitoolbox.torchtrain.train_loop.components.pred_collate_fns import append_predictions, torch_cat_transf class TrainLoopCheckpoint(TrainLoop): def __init__(self, model, train_loader, validation_loader, test_loader, optimizer, criterion, project_name, experiment_name, local_model_result_folder_path, hyperparams, cloud_save_mode='s3', bucket_name='model-result', cloud_dir_prefix='', source_dirs=(), rm_subopt_local_models=False, num_best_checkpoints_kept=2, iteration_save_freq=0, collate_batch_pred_fn=append_predictions, pred_transform_fn=torch_cat_transf, end_auto_eval=True, lazy_experiment_save=False, gpu_mode='single', cuda_device_idx=None, use_amp=False): """TrainLoop with the automatic model check-pointing at the end of each epoch Args: model (TTModel or ModelWrap or TTDataParallel): neural network model train_loader (torch.utils.data.DataLoader): data loader for train data set validation_loader (torch.utils.data.DataLoader or None): data loader for validation data set test_loader (torch.utils.data.DataLoader or None): data loader for test data set optimizer (torch.optim.optimizer.Optimizer or MultiOptimizer): optimizer algorithm. criterion (torch.nn.modules.loss._Loss or MultiLoss or None): criterion during the training procedure project_name (str): root name of the project experiment_name (str): name of the particular experiment local_model_result_folder_path (str): root local path where project folder will be created hyperparams (dict): used hyper-parameters. When running the TrainLoop from jupyter notebook in order to ensure the python experiment file copying to the experiment folder, the user needs to manually specify the python file path as the value for the `experiment_file_path` key. If running the training directly from the terminal the path deduction is done automatically. cloud_save_mode (str or None): Storage destination selector. For AWS S3: 's3' / 'aws_s3' / 'aws' For Google Cloud Storage: 'gcs' / 'google_storage' / 'google storage' Everything else results just in local storage to disk bucket_name (str): name of the bucket in the cloud storage cloud_dir_prefix (str): path to the folder inside the bucket where the experiments are going to be saved source_dirs (list or tuple): paths to the local folders with the source code files used in experiment rm_subopt_local_models (bool or str): if True, the deciding metric is set to 'loss'. Give string metric name to set it as a deciding metric for suboptimal model removal. If metric name consists of substring 'loss' the metric minimization is done otherwise metric maximization is done num_best_checkpoints_kept (int): number of best performing models which are kept when removing suboptimal model checkpoints iteration_save_freq (int): frequency of saving the model checkpoint every specified number of training iterations collate_batch_pred_fn (callable): collate function transforming batch predictions as they come out from the model pred_transform_fn (callable): function transforming all the produced predictions after all the batches have been run through the model end_auto_eval (bool or int): used to optionally disable otherwise automatic end of epoch/training val/test loss calculations. This is useful when conducting very costly experiments to save on compute time. Specify either True/False boolean to always run or never run after each epoch or specify an int to execute only every specified number of epochs. lazy_experiment_save (bool): when in lazy mode experiment tracking components will create the experiment folder only after some training results are available (possibly at the end of the first epoch) instead of at the beginning of training. gpu_mode (str): GPU training mode selection. TrainLoop supports different GPU training modes by specifying one of the following: * ``'single'``: single GPU training * ``'dp'``: multi-GPU training via DataParallel * ``'ddp'``: multi-GPU training via DistributedDataParallel cuda_device_idx (int or None): CUDA device index used when training on multiple GPUs use_amp (bool or dict): use 16-bit Automatic Mixed Precision (AMP) To switch to AMP mode either: * set this parameter to ``True`` to use default AMP ``torch.cuda.amp.GradScaler`` initialization params * provide custom AMP ``torch.cuda.amp.GradScaler`` initialization parameters as a dict as this parameter """ TrainLoop.__init__(self, model, train_loader, validation_loader, test_loader, optimizer, criterion, collate_batch_pred_fn, pred_transform_fn, end_auto_eval, lazy_experiment_save, gpu_mode, cuda_device_idx, use_amp) self.project_name = project_name self.experiment_name = experiment_name self.local_model_result_folder_path = os.path.expanduser(local_model_result_folder_path) self.hyperparams = hyperparams self.cloud_save_mode = cloud_save_mode self.bucket_name = bucket_name self.cloud_dir_prefix = cloud_dir_prefix self.source_dirs = source_dirs self.rm_subopt_local_models = rm_subopt_local_models self.iteration_save_freq = iteration_save_freq if 'experiment_file_path' not in self.hyperparams: self.hyperparams['experiment_file_path'] = inspect.getframeinfo(inspect.currentframe().f_back).filename if 'source_dirs_paths' not in self.hyperparams: self.hyperparams['source_dirs_paths'] = source_dirs if iteration_save_freq == 0: model_checkpoint_cb = ModelCheckpoint( self.project_name, self.experiment_name, self.local_model_result_folder_path, self.hyperparams, cloud_save_mode=self.cloud_save_mode, bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix, rm_subopt_local_models=self.rm_subopt_local_models, num_best_checkpoints_kept=num_best_checkpoints_kept ) elif iteration_save_freq > 0: model_checkpoint_cb = ModelIterationCheckpoint( iteration_save_freq, self.project_name, self.experiment_name, self.local_model_result_folder_path, self.hyperparams, cloud_save_mode=self.cloud_save_mode, bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix, rm_subopt_local_models=self.rm_subopt_local_models, num_best_checkpoints_kept=num_best_checkpoints_kept ) else: raise ValueError('iteration_save_freq can have values only >= 0. ' f'But received value {iteration_save_freq}.') self.callbacks_handler.register_callbacks([model_checkpoint_cb], cache_callbacks=True) class TrainLoopEndSave(TrainLoop): def __init__(self, model, train_loader, validation_loader, test_loader, optimizer, criterion, project_name, experiment_name, local_model_result_folder_path, hyperparams, val_result_package=None, test_result_package=None, cloud_save_mode='s3', bucket_name='model-result', cloud_dir_prefix='', source_dirs=(), collate_batch_pred_fn=append_predictions, pred_transform_fn=torch_cat_transf, end_auto_eval=True, lazy_experiment_save=False, gpu_mode='single', cuda_device_idx=None, use_amp=False): """TrainLoop with the model performance evaluation and final model saving at the end of the training process Args: model (TTModel or ModelWrap or TTDataParallel): neural network model train_loader (torch.utils.data.DataLoader): data loader for train data set validation_loader (torch.utils.data.DataLoader or None): data loader for validation data set test_loader (torch.utils.data.DataLoader or None): data loader for test data set optimizer (torch.optim.optimizer.Optimizer or MultiOptimizer): optimizer algorithm. criterion (torch.nn.modules.loss._Loss or MultiLoss or None): criterion during the training procedure project_name (str): root name of the project experiment_name (str): name of the particular experiment local_model_result_folder_path (str): root local path where project folder will be created hyperparams (dict): used hyper-parameters. When running the TrainLoop from jupyter notebook in order to ensure the python experiment file copying to the experiment folder, the user needs to manually specify the python file path as the value for the `experiment_file_path` key. If running the training directly from the terminal the path deduction is done automatically. val_result_package (AbstractResultPackage or None): result package evaluated on validation data at the end of the training test_result_package (AbstractResultPackage or None): result package evaluated on test data at the end of the training cloud_save_mode (str or None): Storage destination selector. For AWS S3: 's3' / 'aws_s3' / 'aws' For Google Cloud Storage: 'gcs' / 'google_storage' / 'google storage' Everything else results just in local storage to disk bucket_name (str): name of the bucket in the cloud storage cloud_dir_prefix (str): path to the folder inside the bucket where the experiments are going to be saved source_dirs (list or tuple): paths to the local folders with the source code files used in experiment collate_batch_pred_fn (callable): collate function transforming batch predictions as they come out from the model pred_transform_fn (callable): function transforming all the produced predictions after all the batches have been run through the model end_auto_eval (bool or int): used to optionally disable otherwise automatic end of epoch/training val/test loss calculations. This is useful when conducting very costly experiments to save on compute time. Specify either True/False boolean to always run or never run after each epoch or specify an int to execute only every specified number of epochs. lazy_experiment_save (bool): when in lazy mode experiment tracking components will create the experiment folder only after some training results are available (possibly at the end of the first epoch) instead of at the beginning of training. gpu_mode (str): GPU training mode selection. TrainLoop supports different GPU training modes by specifying one of the following: * ``'single'``: single GPU training * ``'dp'``: multi-GPU training via DataParallel * ``'ddp'``: multi-GPU training via DistributedDataParallel cuda_device_idx (int or None): CUDA device index used when training on multiple GPUs use_amp (bool or dict): use 16-bit Automatic Mixed Precision (AMP) To switch to AMP mode either: * set this parameter to ``True`` to use default AMP ``torch.cuda.amp.GradScaler`` initialization params * provide custom AMP ``torch.cuda.amp.GradScaler`` initialization parameters as a dict as this parameter """ TrainLoop.__init__(self, model, train_loader, validation_loader, test_loader, optimizer, criterion, collate_batch_pred_fn, pred_transform_fn, end_auto_eval, lazy_experiment_save, gpu_mode, cuda_device_idx, use_amp) self.project_name = project_name self.experiment_name = experiment_name self.local_model_result_folder_path = os.path.expanduser(local_model_result_folder_path) self.hyperparams = hyperparams self.val_result_package = val_result_package self.test_result_package = test_result_package self.cloud_save_mode = cloud_save_mode self.bucket_name = bucket_name self.cloud_dir_prefix = cloud_dir_prefix self.source_dirs = source_dirs if 'experiment_file_path' not in self.hyperparams: self.hyperparams['experiment_file_path'] = inspect.getframeinfo(inspect.currentframe().f_back).filename if 'source_dirs_paths' not in self.hyperparams: self.hyperparams['source_dirs_paths'] = source_dirs self.check_if_result_packages_possible() self.callbacks_handler.register_callbacks([ ModelTrainEndSave(self.project_name, self.experiment_name, self.local_model_result_folder_path, self.hyperparams, self.val_result_package, self.test_result_package, cloud_save_mode=self.cloud_save_mode, bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix) ], cache_callbacks=True) def check_if_result_packages_possible(self): if self.val_result_package is not None and self.validation_loader is None: raise ValueError('Given the val_result_package but not supplied the validation_loader. ' 'If you want to calculate the val_result_package the validation_loader has to be provided.') if self.test_result_package is not None and self.test_loader is None: raise ValueError('Given the test_result_package but not supplied the test_loader. ' 'If you want to calculate the test_result_package the test_loader has to be provided.') if self.val_result_package is None and self.test_result_package is None: raise ValueError('Both val_result_package and test_result_package are None. ' 'At least one of these should be not None but actual result package.') if self.val_result_package is not None and not isinstance(self.val_result_package, AbstractResultPackage): raise TypeError(f'val_result_package {self.val_result_package} is not inherited from AbstractResultPackage') if self.test_result_package is not None and not isinstance(self.test_result_package, AbstractResultPackage): raise TypeError(f'test_result_package {self.test_result_package} is not inherited from AbstractResultPackage') class TrainLoopCheckpointEndSave(TrainLoopEndSave): def __init__(self, model, train_loader, validation_loader, test_loader, optimizer, criterion, project_name, experiment_name, local_model_result_folder_path, hyperparams, val_result_package=None, test_result_package=None, cloud_save_mode='s3', bucket_name='model-result', cloud_dir_prefix='', source_dirs=(), rm_subopt_local_models=False, num_best_checkpoints_kept=2, iteration_save_freq=0, collate_batch_pred_fn=append_predictions, pred_transform_fn=torch_cat_transf, end_auto_eval=True, lazy_experiment_save=False, gpu_mode='single', cuda_device_idx=None, use_amp=False): """TrainLoop both saving model check-pointing at the end of each epoch and model performance reporting and model saving at the end of the training process Args: model (TTModel or ModelWrap or TTDataParallel): neural network model train_loader (torch.utils.data.DataLoader): data loader for train data set validation_loader (torch.utils.data.DataLoader or None): data loader for validation data set test_loader (torch.utils.data.DataLoader or None): data loader for test data set optimizer (torch.optim.optimizer.Optimizer or MultiOptimizer): optimizer algorithm. criterion (torch.nn.modules.loss._Loss or MultiLoss or None): criterion during the training procedure project_name (str): root name of the project experiment_name (str): name of the particular experiment local_model_result_folder_path (str): root local path where project folder will be created hyperparams (dict): used hyper-parameters. When running the TrainLoop from jupyter notebook in order to ensure the python experiment file copying to the experiment folder, the user needs to manually specify the python file path as the value for the `experiment_file_path` key. If running the training directly from the terminal the path deduction is done automatically. val_result_package (AbstractResultPackage or None): result package evaluated on validation data at the end of the training test_result_package (AbstractResultPackage or None): result package evaluated on test data at the end of the training cloud_save_mode (str or None): Storage destination selector. For AWS S3: 's3' / 'aws_s3' / 'aws' For Google Cloud Storage: 'gcs' / 'google_storage' / 'google storage' Everything else results just in local storage to disk bucket_name (str): name of the bucket in the cloud storage cloud_dir_prefix (str): path to the folder inside the bucket where the experiments are going to be saved source_dirs (list or tuple): paths to the local folders with the source code files used in experiment rm_subopt_local_models (bool or str): if True, the deciding metric is set to 'loss'. Give string metric name to set it as a deciding metric for suboptimal model removal. If metric name consists of substring 'loss' the metric minimization is done otherwise metric maximization is done num_best_checkpoints_kept (int): number of best performing models which are kept when removing suboptimal model checkpoints iteration_save_freq (int): frequency of saving the model checkpoint every specified number of training iterations collate_batch_pred_fn (callable): collate function transforming batch predictions as they come out from the model pred_transform_fn (callable): function transforming all the produced predictions after all the batches have been run through the model end_auto_eval (bool or int): used to optionally disable otherwise automatic end of epoch/training val/test loss calculations. This is useful when conducting very costly experiments to save on compute time. Specify either True/False boolean to always run or never run after each epoch or specify an int to execute only every specified number of epochs. lazy_experiment_save (bool): when in lazy mode experiment tracking components will create the experiment folder only after some training results are available (possibly at the end of the first epoch) instead of at the beginning of training. gpu_mode (str): GPU training mode selection. TrainLoop supports different GPU training modes by specifying one of the following: * ``'single'``: single GPU training * ``'dp'``: multi-GPU training via DataParallel * ``'ddp'``: multi-GPU training via DistributedDataParallel cuda_device_idx (int or None): CUDA device index used when training on multiple GPUs use_amp (bool or dict): use 16-bit Automatic Mixed Precision (AMP) To switch to AMP mode either: * set this parameter to ``True`` to use default AMP ``torch.cuda.amp.GradScaler`` initialization params * provide custom AMP ``torch.cuda.amp.GradScaler`` initialization parameters as a dict as this parameter """ if 'experiment_file_path' not in hyperparams: hyperparams['experiment_file_path'] = inspect.getframeinfo(inspect.currentframe().f_back).filename if 'source_dirs_paths' not in hyperparams: hyperparams['source_dirs_paths'] = source_dirs TrainLoopEndSave.__init__(self, model, train_loader, validation_loader, test_loader, optimizer, criterion, project_name, experiment_name, os.path.expanduser(local_model_result_folder_path), hyperparams, val_result_package, test_result_package, cloud_save_mode, bucket_name, cloud_dir_prefix, source_dirs, collate_batch_pred_fn, pred_transform_fn, end_auto_eval, lazy_experiment_save, gpu_mode, cuda_device_idx, use_amp) self.rm_subopt_local_models = rm_subopt_local_models self.iteration_save_freq = iteration_save_freq if iteration_save_freq == 0: model_checkpoint_cb = ModelCheckpoint( self.project_name, self.experiment_name, self.local_model_result_folder_path, self.hyperparams, cloud_save_mode=self.cloud_save_mode, bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix, rm_subopt_local_models=self.rm_subopt_local_models, num_best_checkpoints_kept=num_best_checkpoints_kept ) elif iteration_save_freq > 0: model_checkpoint_cb = ModelIterationCheckpoint( iteration_save_freq, self.project_name, self.experiment_name, self.local_model_result_folder_path, self.hyperparams, cloud_save_mode=self.cloud_save_mode, bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix, rm_subopt_local_models=self.rm_subopt_local_models, num_best_checkpoints_kept=num_best_checkpoints_kept ) else: raise ValueError('iteration_save_freq can have values only >= 0. ' f'But received value {iteration_save_freq}.') self.callbacks_handler.register_callbacks([model_checkpoint_cb], cache_callbacks=True)
67.83046
122
0.679178
2,933
23,605
5.237982
0.105694
0.033848
0.01777
0.022912
0.948448
0.937577
0.914079
0.890516
0.889475
0.86845
0
0.001627
0.270832
23,605
347
123
68.025937
0.890948
0.505994
0
0.719745
0
0
0.10234
0.01314
0
0
0
0
0
1
0.025478
false
0
0.038217
0
0.082803
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
d12137748bd2c1bf0580f4cfbbb752d0099513d3
853,073
py
Python
rmgpy/test_data/testing_database/thermo/groups/group.py
pw0908/RMG-Py
3846fcce701f2a5fd12dbfa429687e9fcd647298
[ "MIT" ]
1
2022-01-24T05:08:32.000Z
2022-01-24T05:08:32.000Z
rmgpy/test_data/testing_database/thermo/groups/group.py
speth/RMG-Py
1d2c2b684580396e984459d9347628a5ceb80e2e
[ "MIT" ]
72
2016-06-06T18:18:49.000Z
2019-11-17T03:21:10.000Z
rmgpy/test_data/testing_database/thermo/groups/group.py
speth/RMG-Py
1d2c2b684580396e984459d9347628a5ceb80e2e
[ "MIT" ]
3
2017-09-22T15:47:37.000Z
2021-12-30T23:51:47.000Z
#!/usr/bin/env python # encoding: utf-8 name = "Functional Group Additivity Values" shortDesc = u"" longDesc = u""" """ entry( index = -1, label = "R", group = """ 1 * R u0 """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1, label = "C", group = """ 1 * C u0 """, thermo = u'Cs-CsCsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 2, label = "Cbf", group = """ 1 * Cbf u0 """, thermo = u'Cbf-CbCbCbf', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 3, label = "Cbf-CbCbCbf", group = """ 1 * Cbf u0 {2,B} {3,B} {4,B} 2 Cb u0 {1,B} 3 Cb u0 {1,B} 4 Cbf u0 {1,B} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.01,3.68,4.2,4.61,5.2,5.7,6.2],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (4.8,'kcal/mol','+|-',0.17), S298 = (-5,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cbf-CbfCbCb STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 4, label = "Cbf-CbCbfCbf", group = """ 1 * Cbf u0 {2,B} {3,B} {4,B} 2 Cb u0 {1,B} 3 Cbf u0 {1,B} 4 Cbf u0 {1,B} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.01,3.68,4.2,4.61,5.2,5.7,6.2],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (3.7,'kcal/mol','+|-',0.3), S298 = (-5,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cbf-CbfCbfCb STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 5, label = "Cbf-CbfCbfCbf", group = """ 1 * Cbf u0 p0 c0 {3,B} {6,B} {2,B} 2 Cbf u0 p0 c0 {4,B} {5,B} {1,B} 3 Cbf u0 p0 c0 {8,B} {9,B} {1,B} 4 Cbf u0 p0 c0 {10,B} {11,B} {2,B} 5 Cbf u0 p0 c0 {13,B} {14,B} {2,B} 6 Cbf u0 p0 c0 {15,B} {16,B} {1,B} 7 C u0 p0 c0 {8,B} {16,B} 8 C u0 p0 c0 {7,B} {3,B} 9 C u0 p0 c0 {3,B} {10,B} 10 C u0 p0 c0 {9,B} {4,B} 11 C u0 p0 c0 {4,B} {12,B} 12 C u0 p0 c0 {11,B} {13,B} 13 C u0 p0 c0 {12,B} {5,B} 14 C u0 p0 c0 {5,B} {15,B} 15 C u0 p0 c0 {14,B} {6,B} 16 C u0 p0 c0 {7,B} {6,B} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2,3.11,3.9,4.42,5,5.3,5.7],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (1.5,'kcal/mol','+|-',0.3), S298 = (1.8,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cbf-CbfCbfCbf STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" The smallest PAH that can have Cbf-CbfCbfCbf is pyrene. Currently the database is restricted that any group with more three Cbf atoms must have all benzene rings explicitly written out. Previously, this node would also match one carbon on Benzo[c]phenanthrene and does not now. Examples from the original source do not include Benzo[c]phenanthrene. """, ) entry( index = 6, label = "Cb", group = """ 1 * Cb u0 """, thermo = u'Cb-Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 7, label = "Cb-H", group = """ 1 * Cb u0 {2,S} 2 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.24,4.44,5.46,6.3,7.54,8.41,9.73],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (3.3,'kcal/mol','+|-',0.11), S298 = (11.53,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cb-H BENSON""", longDesc = u""" """, ) entry( index = 8, label = "Cb-O2s", group = """ 1 * Cb u0 {2,S} 2 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.9,5.3,6.2,6.6,6.9,6.9,7.07],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-0.9,'kcal/mol','+|-',0.16), S298 = (-10.2,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cb-O BENSON Cp1500=3D Cp1000*(Cp1500/Cp1000: Cb/Cd)""", longDesc = u""" """, ) entry( index = 1197, label = "Cb-S2s", group = """ 1 * Cb u0 {2,S} 2 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2.46,3.24,3.92,4.49,5.27,5.75,6.3],'cal/(mol*K)'), H298 = (5.83,'kcal/mol'), S298 = (-7.94,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 9, label = "Cb-C", group = """ 1 * Cb u0 {2,S} 2 C u0 {1,S} """, thermo = u'Cb-Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 10, label = "Cb-Cs", group = """ 1 * Cb u0 {2,S} 2 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2.67,3.14,3.68,4.15,4.96,5.44,5.98],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (5.51,'kcal/mol','+|-',0.13), S298 = (-7.69,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cb-Cs BENSON""", longDesc = u""" """, ) entry( index = 11, label = "Cb-Cds", group = """ 1 * Cb u0 {2,S} 2 [Cd,CO] u0 {1,S} """, thermo = u'Cb-(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 12, label = "Cb-(Cds-O2d)", group = """ 1 * Cb u0 {2,S} 2 CO u0 {1,S} {3,D} 3 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.59,3.97,4.38,4.72,5.28,5.61,5.75],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (3.69,'kcal/mol','+|-',0.2), S298 = (-7.8,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Enthalpy from Cb-CO, entropies and heat capacities assigned value of Cb-Cd""", longDesc = u""" """, ) entry( index = 13, label = "Cb-(Cds-Cd)", group = """ 1 * Cb u0 {2,S} 2 Cd u0 {1,S} {3,D} 3 C u0 {2,D} """, thermo = u'Cb-(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 14, label = "Cb-(Cds-Cds)", group = """ 1 * Cb u0 {2,S} 2 Cd u0 {1,S} {3,D} 3 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.59,3.97,4.38,4.72,5.28,5.61,5.75],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (5.69,'kcal/mol','+|-',0.2), S298 = (-7.8,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cb-Cd STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 15, label = "Cb-(Cds-Cdd)", group = """ 1 * Cb u0 {2,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} """, thermo = u'Cb-(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 16, label = "Cb-(Cds-Cdd-O2d)", group = """ 1 * Cb u0 {2,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {4,D} 4 O2d u0 {3,D} """, thermo = u'Cb-(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cb-(Cds-Cdd-S2d)", group = """ 1 * Cb u0 {2,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {4,D} 4 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 17, label = "Cb-(Cds-Cdd-Cd)", group = """ 1 * Cb u0 {2,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {4,D} 4 C u0 {3,D} """, thermo = u'Cb-(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 18, label = "Cb-Ct", group = """ 1 * Cb u0 {2,S} 2 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.59,3.97,4.38,4.72,5.28,5.61,5.75],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (5.69,'kcal/mol','+|-',0.3), S298 = (-7.8,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cb-Ct STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 1839, label = "Cb-(CtN3t)", group = """ 1 * Cb u0 {2,S} 2 Ct u0 {1,S} {3,T} 3 N3t u0 {2,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([9.8,11.2,12.3,13.1,14.2,14.9,16.65],'cal/(mol*K)'), H298 = (35.8,'kcal/mol'), S298 = (20.5,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 19, label = "Cb-Cb", group = """ 1 * Cb u0 {2,S} 2 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.33,4.22,4.89,5.27,5.76,5.95,6.05],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (4.96,'kcal/mol','+|-',0.3), S298 = (-8.64,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cb-Cb BENSON""", longDesc = u""" """, ) entry( index = 1821, label = "Cb-N3s", group = """ 1 * Cb u0 {2,S} 2 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.95,5.21,5.94,6.32,6.53,6.56,6.635],'cal/(mol*K)'), H298 = (-0.5,'kcal/mol'), S298 = (-9.69,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1198, label = "Cb-C=S", group = """ 1 * Cb u0 {2,S} 2 CS u0 {1,S} {3,D} 3 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2.46,3.24,3.92,4.49,5.27,5.75,6.3],'cal/(mol*K)'), H298 = (5.83,'kcal/mol'), S298 = (-7.94,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 20, label = "Ct", group = """ 1 * Ct u0 """, thermo = u'Ct-CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1849, label = "Ct-CtN3s", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 N3s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1853, label = "Ct-N3tN3s", group = """ 1 * Ct u0 {2,T} {3,S} 2 N3t u0 {1,T} 3 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 21, label = "Ct-CtH", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.28,5.99,6.49,6.87,7.47,7.96,8.85],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (26.93,'kcal/mol','+|-',0.05), S298 = (24.7,'cal/(mol*K)','+|-',0.07), ), shortDesc = u"""Ct-H STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 22, label = "Ct-CtOs", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.64,4.39,4.85,5.63,5.66,5.73,5.73],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (31.4,'kcal/mol','+|-',0.27), S298 = (4.91,'cal/(mol*K)','+|-',0.09), ), shortDesc = u"""Ct-O MELIUS / hc#coh !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1852, label = "Ct-N3tOs", group = """ 1 * Ct u0 {2,T} {3,S} 2 N3t u0 {1,T} 3 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1195, label = "Ct-CtSs", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.29,3.67,4,4.29,4.74,5.05,5.49],'cal/(mol*K)'), H298 = (27.63,'kcal/mol'), S298 = (6.32,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1942, label = "Ct-N3tC", group = """ 1 * Ct u0 {2,T} {3,S} 2 N3t u0 {1,T} 3 C u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1850, label = "Ct-N3tCs", group = """ 1 * Ct u0 {2,T} {3,S} 2 N3t u0 {1,T} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1851, label = "Ct-N3tCd", group = """ 1 * Ct u0 {2,T} {3,S} 2 N3t u0 {1,T} 3 Cd u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 23, label = "Ct-CtC", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 C u0 {1,S} """, thermo = u'Ct-CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 24, label = "Ct-CtCs", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.13,3.48,3.81,4.09,4.6,4.92,6.35],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (27.55,'kcal/mol','+|-',0.27), S298 = (6.35,'cal/(mol*K)','+|-',0.09), ), shortDesc = u"""Ct-Cs STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 25, label = "Ct-CtCds", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 [Cd,CO] u0 {1,S} """, thermo = u'Ct-Ct(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 26, label = "Ct-Ct(Cds-O2d)", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 CO u0 {1,S} {4,D} 4 O2d u0 {3,D} """, thermo = u'Ct-CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 27, label = "Ct-Ct(Cds-Cd)", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Cd u0 {1,S} {4,D} 4 C u0 {3,D} """, thermo = u'Ct-Ct(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 28, label = "Ct-Ct(Cds-Cds)", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Cd u0 {1,S} {4,D} 4 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2.57,3.54,3.5,4.92,5.34,5.5,5.8],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (28.2,'kcal/mol','+|-',0.27), S298 = (6.43,'cal/(mol*K)','+|-',0.09), ), shortDesc = u"""Ct-Cd STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 29, label = "Ct-Ct(Cds-Cdd)", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Cd u0 {1,S} {4,D} 4 Cdd u0 {3,D} """, thermo = u'Ct-Ct(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 30, label = "Ct-Ct(Cds-Cdd-O2d)", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Cd u0 {1,S} {4,D} 4 Cdd u0 {3,D} {5,D} 5 O2d u0 {4,D} """, thermo = u'Ct-Ct(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Ct-Ct(Cds-Cdd-S2d)", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Cd u0 {1,S} {4,D} 4 Cdd u0 {3,D} {5,D} 5 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 31, label = "Ct-Ct(Cds-Cdd-Cd)", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Cd u0 {1,S} {4,D} 4 Cdd u0 {3,D} {5,D} 5 C u0 {4,D} """, thermo = u'Ct-Ct(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 32, label = "Ct-CtCt", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.54,4.06,4.4,4.64,5,5.23,5.57],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (25.6,'kcal/mol','+|-',0.27), S298 = (5.88,'cal/(mol*K)','+|-',0.09), ), shortDesc = u"""Ct-Ct STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 1840, label = "Ct-Ct(CtN3t)", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Ct u0 {1,S} {4,T} 4 N3t u0 {3,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.3,11.3,12.1,12.7,13.6,14.3,15.3],'cal/(mol*K)'), H298 = (63.8,'kcal/mol'), S298 = (35.4,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 33, label = "Ct-CtCb", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2.57,3.54,4.5,4.92,5.34,5.5,5.8],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (24.67,'kcal/mol','+|-',0.27), S298 = (6.43,'cal/(mol*K)','+|-',0.09), ), shortDesc = u"""Ct-Cb STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""", longDesc = u""" """, ) entry( index = 1196, label = "Ct-CtC=S", group = """ 1 * Ct u0 {2,T} {3,S} 2 Ct u0 {1,T} 3 CS u0 {1,S} {4,D} 4 S2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.29,3.67,4,4.29,4.74,5.05,5.49],'cal/(mol*K)'), H298 = (27.63,'kcal/mol'), S298 = (6.32,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 34, label = "Cdd", group = """ 1 * Cdd u0 """, thermo = u'Cdd-CdsCds', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1854, label = "Cdd-N3dCd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 N3d u0 {1,D} 3 Cd u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.5,5.1,5.6,6,6.5,6.9,7.4],'cal/(mol*K)','+|-',[1.1,1.1,1.1,1.1,1.1,1.1,1.1]), H298 = (25.9,'kcal/mol','+|-',1.5), S298 = (19.7,'cal/(mol*K)','+|-',1.4), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 35, label = "Cdd-OdOd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 O2d u0 {1,D} 3 O2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.91,9.86,10.65,11.31,12.32,12.99,13.93],'cal/(mol*K)'), H298 = (-94.05,'kcal/mol','+|-',0.03), S298 = (52.46,'cal/(mol*K)','+|-',0.002), ), shortDesc = u"""CHEMKIN DATABASE: S(group) = S(CO2) + Rln(2)""", longDesc = u""" """, ) entry( index = 1466, label = "Cdd-OdSd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 O2d u0 {1,D} 3 S2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([9.81,10.8,11.6,12.21,13.03,13.51,14.12],'cal/(mol*K)'), H298 = (-35.96,'kcal/mol'), S298 = (55.34,'cal/(mol*K)'), ), shortDesc = u"""CAC calc 1D-HR""", longDesc = u""" """, ) entry( index = 1199, label = "Cdd-SdSd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 S2d u0 {1,D} 3 S2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.91,11.83,12.49,12.98,13.63,14.01,14.47],'cal/(mol*K)'), H298 = (24.5,'kcal/mol'), S298 = (58.24,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2009""", longDesc = u""" """, ) entry( index = 36, label = "Cdd-CdOd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 C u0 {1,D} 3 O2d u0 {1,D} """, thermo = u'Cdd-CdsOd', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 37, label = "Cdd-CdsOd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cd u0 {1,D} 3 O2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""O=C*=C< currently treat the adjacent C as Ck""", longDesc = u""" """, ) entry( index = 38, label = "Cdd-CddOd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} 3 O2d u0 {1,D} """, thermo = u'Cdd-(Cdd-Cd)O2d', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 40, label = "Cdd-(Cdd-O2d)O2d", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 O2d u0 {1,D} 4 O2d u0 {2,D} """, thermo = u'Cdd-CdsOd', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 39, label = "Cdd-(Cdd-Cd)O2d", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 O2d u0 {1,D} 4 C u0 {2,D} """, thermo = u'Cdd-CdsOd', shortDesc = u"""O=C*=C= currently not defined. Assigned same value as Ca""", longDesc = u""" """, ) entry( index = 1200, label = "Cdd-CdSd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 C u0 {1,D} 3 S2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.88,8.48,8.8,8.99,9.23,9.37,9.58],'cal/(mol*K)'), H298 = (40.33,'kcal/mol'), S298 = (34.24,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cdd-CdsSd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cd u0 {1,D} 3 S2d u0 {1,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cdd-CddSd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} 3 S2d u0 {1,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cdd-(Cdd-S2d)S2d", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 S2d u0 {1,D} 4 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cdd-(Cdd-Cd)S2d", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 S2d u0 {1,D} 4 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 41, label = "Cdd-CdCd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 C u0 {1,D} 3 C u0 {1,D} """, thermo = u'Cdd-CdsCds', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 42, label = "Cdd-CddCdd", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} 3 Cdd u0 {1,D} """, thermo = u'Cdd-(Cdd-Cd)(Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 43, label = "Cdd-(Cdd-O2d)(Cdd-O2d)", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 Cdd u0 {1,D} {5,D} 4 O2d u0 {2,D} 5 O2d u0 {3,D} """, thermo = u'Cdd-CdsCds', shortDesc = u"""O=C=C*=C=O, currently not defined. Assigned same value as Ca""", longDesc = u""" """, ) entry( index = -1, label = "Cdd-(Cdd-S2d)(Cdd-S2d)", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 Cdd u0 {1,D} {5,D} 4 S2d u0 {2,D} 5 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 44, label = "Cdd-(Cdd-O2d)(Cdd-Cd)", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 Cdd u0 {1,D} {5,D} 4 O2d u0 {2,D} 5 C u0 {3,D} """, thermo = u'Cdd-(Cdd-O2d)Cds', shortDesc = u"""O=C=C*=C=C, currently not defined. Assigned same value as Ca""", longDesc = u""" """, ) entry( index = -1, label = "Cdd-(Cdd-S2d)(Cdd-Cd)", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 Cdd u0 {1,D} {5,D} 4 S2d u0 {2,D} 5 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 45, label = "Cdd-(Cdd-Cd)(Cdd-Cd)", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 Cdd u0 {1,D} {5,D} 4 C u0 {2,D} 5 C u0 {3,D} """, thermo = u'Cdd-CdsCds', shortDesc = u"""C=C=C*=C=C, currently not defined. Assigned same value as Ca""", longDesc = u""" """, ) entry( index = 46, label = "Cdd-CddCds", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} 3 Cd u0 {1,D} """, thermo = u'Cdd-(Cdd-Cd)(Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 47, label = "Cdd-(Cdd-O2d)Cds", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 Cd u0 {1,D} 4 O2d u0 {2,D} """, thermo = u'Cdd-CdsCds', shortDesc = u"""O=C=C*=C<, currently not defined. Assigned same value as Ca """, longDesc = u""" """, ) entry( index = -1, label = "Cdd-(Cdd-S2d)Cds", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 Cd u0 {1,D} 4 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 48, label = "Cdd-(Cdd-Cd)Cds", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cdd u0 {1,D} {4,D} 3 Cd u0 {1,D} 4 C u0 {2,D} """, thermo = u'Cdd-CdsCds', shortDesc = u"""C=C=C*=C<, currently not defined. Assigned same value as Ca """, longDesc = u""" """, ) entry( index = 49, label = "Cdd-CdsCds", group = """ 1 * Cdd u0 {2,D} {3,D} 2 Cd u0 {1,D} 3 Cd u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.9,4.4,4.7,5,5.3,5.5,5.7],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (34.2,'kcal/mol','+|-',0.2), S298 = (6,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Benson's Ca """, longDesc = u""" """, ) entry( index = 50, label = "Cds", group = """ 1 * [Cd,CO,CS] u0 """, thermo = u'Cds-CdsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1823, label = "Cds-OdN3sH", group = """ 1 * CO u0 {2,S} {3,S} {4,D} 2 N3s u0 {1,S} 3 H u0 {1,S} 4 O2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.03,7.87,8.82,9.68,11.16,12.2,14.8],'cal/(mol*K)'), H298 = (-29.6,'kcal/mol'), S298 = (34.93,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1824, label = "Cds-OdN3sCs", group = """ 1 * CO u0 {2,S} {3,S} {4,D} 2 N3s u0 {1,S} 3 Cs u0 {1,S} 4 O2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.37,6.17,7.07,7.66,9.62,11.19,15.115],'cal/(mol*K)'), H298 = (-32.8,'kcal/mol'), S298 = (16.2,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1855, label = "Cd-N3dCsCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 N3d u0 {1,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.5,4.2,5,5.6,6.6,7.2,7.9],'cal/(mol*K)','+|-',[0.9,0.9,0.9,0.9,0.9,0.9,0.9]), H298 = (5.7,'kcal/mol','+|-',1.2), S298 = (2,'cal/(mol*K)','+|-',1.1), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1856, label = "Cd-N3dCsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 N3d u0 {1,D} 3 Cs u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.5,6.3,7.2,8,9.3,10.2,11.6],'cal/(mol*K)','+|-',[0.9,0.9,0.9,0.9,0.9,0.9,0.9]), H298 = (3.3,'kcal/mol','+|-',1.3), S298 = (21.2,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1857, label = "Cd-N3dHH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 N3d u0 {1,D} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.2,7.4,8.7,9.8,11.5,12.9,15],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (4.4,'kcal/mol','+|-',1.4), S298 = (40.8,'cal/(mol*K)','+|-',1.3), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 51, label = "Cds-OdHH", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.47,9.38,10.46,11.52,13.37,14.81,14.81],'cal/(mol*K)','+|-',[0.06,0.06,0.06,0.06,0.06,0.06,0.06]), H298 = (-25.95,'kcal/mol','+|-',0.11), S298 = (53.68,'cal/(mol*K)','+|-',0.06), ), shortDesc = u"""CO-HH BENSON !!!WARNING! Cp1500 value taken as Cp1000, S(group) = S(CH2O) + Rln(2)""", longDesc = u""" """, ) entry( index = 52, label = "Cds-OdOsH", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 O2s u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([25.88,34.56,42.08,48.16,56.57,61.38,65.84],'J/(mol*K)','+|-',[4.01,4.01,4.01,4.01,4.01,4.01,4.01]), H298 = (-211.8,'kJ/mol','+|-',3.42), S298 = (124.04,'J/(mol*K)','+|-',4.68), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1454, label = "CO-SsH", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 S2s u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.51,6.16,6.7,7.17,8.06,8.79,9.83],'cal/(mol*K)'), H298 = (-9.84,'kcal/mol'), S298 = (29.36,'cal/(mol*K)'), ), shortDesc = u"""CAC 1d-HR calc""", longDesc = u""" """, ) entry( index = 53, label = "Cds-OdOsOs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([26.17,39.3,48.25,53.88,58.97,59.63,56.09],'J/(mol*K)','+|-',[6,6,6,6,6,6,6]), H298 = (-281.4,'kJ/mol','+|-',5.11), S298 = (22.66,'J/(mol*K)','+|-',7), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1455, label = "CO-CsSs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 S2s u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.37,5.04,5.51,5.83,6.29,6.48,6.38],'cal/(mol*K)'), H298 = (-14.02,'kcal/mol'), S298 = (8.55,'cal/(mol*K)'), ), shortDesc = u"""CAC 1d-HR calc""", longDesc = u""" """, ) entry( index = 1456, label = "CO-OsSs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 O2s u0 {1,S} 4 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.94,5.63,6.35,6.99,7.59,7.76,8.18],'cal/(mol*K)'), H298 = (-11.53,'kcal/mol'), S298 = (9.61,'cal/(mol*K)'), ), shortDesc = u"""CAC CBS-QB3 1Dhr calc""", longDesc = u""" """, ) entry( index = 54, label = "Cds-OdCH", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 C u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-OdCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 55, label = "Cds-OdCsH", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cs u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([26.24,31.22,35.94,40.13,46.74,51.39,57.73],'J/(mol*K)','+|-',[2.08,2.08,2.08,2.08,2.08,2.08,2.08]), H298 = (-123.4,'kJ/mol','+|-',1.77), S298 = (145.46,'J/(mol*K)','+|-',2.42), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 56, label = "Cds-OdCdsH", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 57, label = "Cds-O2d(Cds-O2d)H", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 CO u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([29.76,34.63,39.25,43.32,49.57,53.77,59.32],'J/(mol*K)','+|-',[1.7,1.7,1.7,1.7,1.7,1.7,1.7]), H298 = (-104.8,'kJ/mol','+|-',1.45), S298 = (140.49,'J/(mol*K)','+|-',1.98), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 58, label = "Cds-O2d(Cds-Cd)H", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 H u0 {1,S} 5 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([27.31,34,39.42,43.77,50.16,54.55,60.77],'J/(mol*K)','+|-',[4.9,4.9,4.9,4.9,4.9,4.9,4.9]), H298 = (-128.3,'kJ/mol','+|-',5.9), S298 = (129.26,'J/(mol*K)','+|-',5.71), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 59, label = "Cds-O2d(Cds-Cds)H", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 H u0 {1,S} 5 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.45,8.77,9.82,10.7,12.14,12.9,12.9],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (-30.9,'kcal/mol','+|-',0.3), S298 = (33.4,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""CO-CdH Hf BOZZELLI S,Cp =3D CO/C/H-del(cd syst) !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 60, label = "Cds-O2d(Cds-Cdd)H", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 H u0 {1,S} 5 Cdd u0 {2,D} """, thermo = u'Cds-O2d(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 61, label = "Cds-O2d(Cds-Cdd-O2d)H", group = """ 1 * CO u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 O2d u0 {1,D} 5 H u0 {1,S} 6 O2d u0 {3,D} """, thermo = u'Cds-O2d(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 62, label = "Cds-O2d(Cds-Cdd-Cd)H", group = """ 1 * CO u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 O2d u0 {1,D} 5 H u0 {1,S} 6 C u0 {3,D} """, thermo = u'Cds-O2d(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 63, label = "Cds-OdCtH", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 64, label = "Cds-OdCbH", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 65, label = "Cds-OdCOs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 C u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-OdCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 66, label = "Cds-OdCsOs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([20.67,28.39,34.6,39.48,46.23,50.09,52.68],'J/(mol*K)','+|-',[2.85,2.85,2.85,2.85,2.85,2.85,2.85]), H298 = (-222,'kJ/mol','+|-',2.43), S298 = (43.52,'J/(mol*K)','+|-',3.33), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 67, label = "Cds-OdCdsOs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 68, label = "Cds-O2d(Cds-O2d)O2s", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 CO u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 O2s u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([27.18,34.34,39.85,44.13,49.81,52.4,52.33],'J/(mol*K)','+|-',[3.36,3.36,3.36,3.36,3.36,3.36,3.36]), H298 = (-196.2,'kJ/mol','+|-',2.86), S298 = (39.37,'J/(mol*K)','+|-',3.92), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 69, label = "Cds-O2d(Cds-Cd)O2s", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 O2s u0 {1,S} 5 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([28.33,37.84,44.54,49.34,55.45,58.73,60.61],'J/(mol*K)','+|-',[7.46,7.46,7.46,7.46,7.46,7.46,7.46]), H298 = (-218.6,'kJ/mol','+|-',6.36), S298 = (33.44,'J/(mol*K)','+|-',8.7), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 70, label = "Cds-O2d(Cds-Cds)O2s", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 O2s u0 {1,S} 5 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.97,6.7,7.4,8.02,8.87,9.36,9.36],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (-32.1,'kcal/mol','+|-',0.3), S298 = (14.78,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""CO-OCd RPS + S Coreected !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 71, label = "Cds-O2d(Cds-Cdd)O2s", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 O2s u0 {1,S} 5 Cdd u0 {2,D} """, thermo = u'Cds-O2d(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 72, label = "Cds-O2d(Cds-Cdd-O2d)O2s", group = """ 1 * CO u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 O2d u0 {1,D} 5 O2s u0 {1,S} 6 O2d u0 {3,D} """, thermo = u'Cds-O2d(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 73, label = "Cds-O2d(Cds-Cdd-Cd)O2s", group = """ 1 * CO u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 O2d u0 {1,D} 5 O2s u0 {1,S} 6 C u0 {3,D} """, thermo = u'Cds-O2d(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 74, label = "Cds-OdCtOs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 75, label = "Cds-OdCbOs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.97,6.7,7.4,8.02,8.87,9.36,9.36],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (-36.6,'kcal/mol','+|-',0.3), S298 = (14.78,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""CO-OCb RPS + S Coreected !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 76, label = "Cds-OdCC", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 C u0 {1,S} 4 C u0 {1,S} """, thermo = u'Cds-OdCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 77, label = "Cds-OdCsCs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([23.82,27.7,31.22,34.19,38.37,40.85,43.26],'J/(mol*K)','+|-',[2.08,2.08,2.08,2.08,2.08,2.08,2.08]), H298 = (-132.2,'kJ/mol','+|-',1.77), S298 = (61.78,'J/(mol*K)','+|-',2.42), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 78, label = "Cds-OdCdsCs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 79, label = "Cds-O2d(Cds-O2d)Cs", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 CO u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 Cs u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([26.77,30.83,34.36,37.27,41.27,43.45,45.25],'J/(mol*K)','+|-',[1.28,1.28,1.28,1.28,1.28,1.28,1.28]), H298 = (-122,'kJ/mol','+|-',1.09), S298 = (57.8,'J/(mol*K)','+|-',1.5), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 80, label = "Cds-O2d(Cds-Cd)Cs", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 Cs u0 {1,S} 5 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([25.26,30.66,34.68,37.69,41.62,43.93,46.69],'J/(mol*K)','+|-',[4.9,4.9,4.9,4.9,4.9,4.9,4.9]), H298 = (-130.4,'kJ/mol','+|-',4.17), S298 = (47.38,'J/(mol*K)','+|-',5.71), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 81, label = "Cds-O2d(Cds-Cds)Cs", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 Cs u0 {1,S} 5 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.46,6.32,7.17,7.88,9,9.77,9.77],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (-30.9,'kcal/mol','+|-',0.3), S298 = (14.6,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""CO-CdCs Hf BENSON =3D CO/Cb/C S,Cp !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 82, label = "Cds-O2d(Cds-Cdd)Cs", group = """ 1 * CO u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 O2d u0 {1,D} 4 Cs u0 {1,S} 5 Cdd u0 {2,D} """, thermo = u'Cds-O2d(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 83, label = "Cds-O2d(Cds-Cdd-O2d)Cs", group = """ 1 * CO u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 O2d u0 {1,D} 5 Cs u0 {1,S} 6 O2d u0 {3,D} """, thermo = u'Cds-O2d(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 84, label = "Cds-O2d(Cds-Cdd-Cd)Cs", group = """ 1 * CO u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 O2d u0 {1,D} 5 Cs u0 {1,S} 6 C u0 {3,D} """, thermo = u'Cds-O2d(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 85, label = "Cds-OdCdsCds", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 86, label = "Cds-O2d(Cds-O2d)(Cds-O2d)", group = """ 1 * CO u0 {2,S} {3,S} {4,D} 2 CO u0 {1,S} {5,D} 3 CO u0 {1,S} {6,D} 4 O2d u0 {1,D} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([31.75,33.35,34.1,34.51,35.19,36.06,38.14],'J/(mol*K)','+|-',[2.41,2.41,2.41,2.41,2.41,2.41,2.41]), H298 = (-89.3,'kJ/mol','+|-',2.05), S298 = (64.51,'J/(mol*K)','+|-',2.81), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 87, label = "Cds-O2d(Cds-Cd)(Cds-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 CO u0 {1,S} {6,D} 5 C u0 {3,D} 6 O2d u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 88, label = "Cds-O2d(Cds-Cds)(Cds-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 CO u0 {1,S} {6,D} 5 Cd u0 {3,D} 6 O2d u0 {4,D} """, thermo = u'Cds-O2d(Cds-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 89, label = "Cds-O2d(Cds-Cdd)(Cds-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 CO u0 {1,S} {6,D} 5 Cdd u0 {3,D} 6 O2d u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cdd-Cd)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 90, label = "Cds-O2d(Cds-Cdd-O2d)(Cds-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 CO u0 {1,S} {7,D} 5 Cdd u0 {3,D} {6,D} 6 O2d u0 {5,D} 7 O2d u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 91, label = "Cds-O2d(Cds-Cdd-Cd)(Cds-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 CO u0 {1,S} {7,D} 5 Cdd u0 {3,D} {6,D} 6 C u0 {5,D} 7 O2d u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 92, label = "Cds-O2d(Cds-Cd)(Cds-Cd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 C u0 {3,D} 6 C u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 93, label = "Cds-O2d(Cds-Cds)(Cds-Cds)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {3,D} 6 Cd u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.46,6.32,7.17,7.88,9,9.77,9.77],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (-30.9,'kcal/mol','+|-',0.3), S298 = (14.6,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""CO-CdCd Estimate BOZZELLI !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 94, label = "Cds-O2d(Cds-Cdd)(Cds-Cds)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 Cdd u0 {3,D} 6 Cd u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cdd-Cd)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 95, label = "Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 Cdd u0 {3,D} {7,D} 6 Cd u0 {4,D} 7 O2d u0 {5,D} """, thermo = u'Cds-O2d(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 96, label = "Cds-O2d(Cds-Cdd-Cd)(Cds-Cds)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 Cdd u0 {3,D} {7,D} 6 Cd u0 {4,D} 7 C u0 {5,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 97, label = "Cds-O2d(Cds-Cdd)(Cds-Cdd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 Cdd u0 {3,D} 6 Cdd u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 98, label = "Cds-O2d(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 Cdd u0 {3,D} {7,D} 6 Cdd u0 {4,D} {8,D} 7 O2d u0 {5,D} 8 O2d u0 {6,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 99, label = "Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 Cdd u0 {3,D} {7,D} 6 Cdd u0 {4,D} {8,D} 7 C u0 {5,D} 8 O2d u0 {6,D} """, thermo = u'Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 100, label = "Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,S} {6,D} 5 Cdd u0 {3,D} {7,D} 6 Cdd u0 {4,D} {8,D} 7 C u0 {5,D} 8 C u0 {6,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 101, label = "Cds-OdCtCs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 102, label = "Cds-OdCtCds", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-OdCt(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 103, label = "Cds-OdCt(Cds-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 CO u0 {1,S} {5,D} 5 O2d u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 104, label = "Cds-OdCt(Cds-Cd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 C u0 {4,D} """, thermo = u'Cds-OdCt(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 105, label = "Cds-OdCt(Cds-Cds)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 Cd u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 106, label = "Cds-OdCt(Cds-Cdd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 Cdd u0 {4,D} """, thermo = u'Cds-OdCt(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 107, label = "Cds-OdCt(Cds-Cdd-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 Cdd u0 {4,D} {6,D} 6 O2d u0 {5,D} """, thermo = u'Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 108, label = "Cds-OdCt(Cds-Cdd-Cd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 Cdd u0 {4,D} {6,D} 6 C u0 {5,D} """, thermo = u'Cds-OdCt(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 109, label = "Cds-OdCtCt", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 110, label = "Cds-OdCbCs", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 111, label = "Cds-OdCbCds", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-OdCb(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 112, label = "Cds-OdCb(Cds-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 CO u0 {1,S} {5,D} 5 O2d u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 113, label = "Cds-OdCb(Cds-Cd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 C u0 {4,D} """, thermo = u'Cds-OdCb(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 114, label = "Cds-OdCb(Cds-Cds)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 Cd u0 {4,D} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 115, label = "Cds-OdCb(Cds-Cdd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 Cdd u0 {4,D} """, thermo = u'Cds-OdCb(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 116, label = "Cds-OdCb(Cds-Cdd-O2d)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 Cdd u0 {4,D} {6,D} 6 O2d u0 {5,D} """, thermo = u'Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 117, label = "Cds-OdCb(Cds-Cdd-Cd)", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 Cd u0 {1,S} {5,D} 5 Cdd u0 {4,D} {6,D} 6 C u0 {5,D} """, thermo = u'Cds-OdCb(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 118, label = "Cds-OdCbCt", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} """, thermo = u'Cds-OdCt(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 119, label = "Cds-OdCbCb", group = """ 1 * CO u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} """, thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 120, label = "Cds-CdHH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-CdsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 121, label = "Cds-CdsHH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.1,6.36,7.51,8.5,10.07,11.27,13.19],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (6.26,'kcal/mol','+|-',0.19), S298 = (27.61,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-HH BENSON""", longDesc = u""" """, ) entry( index = 122, label = "Cds-CddHH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 123, label = "Cds-(Cdd-O2d)HH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.08,13.91,15.4,16.64,18.61,20.1,22.47],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (-11.34,'kcal/mol','+|-',0.19), S298 = (57.47,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{CCO/H2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)HH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 124, label = "Cds-(Cdd-Cd)HH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 125, label = "Cds-CdOsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 O2s u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([18.08,21.17,24.43,27.41,32.22,35.73,40.97],'J/(mol*K)'), H298 = (36.4,'kJ/mol'), S298 = (33.51,'J/(mol*K)'), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 126, label = "Cds-CdsOsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 O2s u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.75,6.46,7.64,8.35,9.1,9.56,10.46],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (2.03,'kcal/mol','+|-',0.19), S298 = (6.2,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-OH BOZZELLI Hf vin-oh RADOM + C/Cd/H, S&Cp LAY""", longDesc = u""" """, ) entry( index = 127, label = "Cds-CddOsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 O2s u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 128, label = "Cds-(Cdd-O2d)OsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.29,13.67,15.1,16.1,17.36,18.25,19.75],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (2.11,'kcal/mol','+|-',0.19), S298 = (38.17,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{CCO/O/H} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 129, label = "Cds-(Cdd-Cd)OsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdSsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 S2s u0 {1,S} 4 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1180, label = "Cds-CdsSsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 S2s u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.41,5.2,5.98,6.68,7.8,8.62,9.84],'cal/(mol*K)'), H298 = (8.87,'kcal/mol'), S298 = (7.87,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CddSsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 S2s u0 {1,S} 4 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)SsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)SsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 130, label = "Cds-CdOsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.34,11.93,14.86,17.95,22.31,24.6,26.92],'J/(mol*K)','+|-',[7.4,7.4,7.4,7.4,7.4,7.4,7.4]), H298 = (28.3,'kJ/mol','+|-',6.3), S298 = (-42.69,'J/(mol*K)','+|-',8.63), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 131, label = "Cds-CdsOsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-CdsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 132, label = "Cds-CddOsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 133, label = "Cds-(Cdd-O2d)OsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.56,15.58,17.69,18.67,18.78,18.4,18.01],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (2.403,'kcal/mol','+|-',0.19), S298 = (13.42,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{CCO/O2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 134, label = "Cds-(Cdd-Cd)OsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdSsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsSsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CddSsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)SsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)SsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 135, label = "Cds-CdCH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 C u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-CdsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 136, label = "Cds-CdsCsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cs u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.16,5.03,5.81,6.5,7.65,8.45,9.62],'cal/(mol*K)','+|-',[0.06,0.06,0.06,0.06,0.06,0.06,0.06]), H298 = (8.59,'kcal/mol','+|-',0.17), S298 = (7.97,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-CsH BENSON""", longDesc = u""" """, ) entry( index = 137, label = "Cds-CdsCdsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-Cds(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 139, label = "Cds-Cds(Cds-Cd)H", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 H u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 140, label = "Cds-Cds(Cds-Cds)H", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 H u0 {1,S} 5 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.46,5.79,6.75,7.42,8.35,8.99,9.98],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (6.78,'kcal/mol','+|-',0.2), S298 = (6.38,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-CdH BENSON""", longDesc = u""" """, ) entry( index = 141, label = "Cds-Cds(Cds-Cdd)H", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 H u0 {1,S} 5 Cdd u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd-S2d)H", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 H u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 143, label = "Cds-Cds(Cds-Cdd-Cd)H", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 H u0 {1,S} 6 C u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 144, label = "Cds-CdsCtH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Ct u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.46,5.79,6.75,7.42,8.35,8.99,9.98],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (6.78,'kcal/mol','+|-',0.2), S298 = (6.38,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-CtH BENSON""", longDesc = u""" """, ) entry( index = 1836, label = "Cds-CdsH(CtN3t)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Ct u0 {1,S} {5,T} 3 Cd u0 {1,D} 4 H u0 {1,S} 5 N3t u0 {2,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.3,12,13.4,14.6,16.3,17.5,19.4],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (38.5,'kcal/mol','+|-',1.3), S298 = (37.6,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 145, label = "Cds-CdsCbH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cb u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.46,5.79,6.75,7.42,8.35,8.99,9.98],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (6.78,'kcal/mol','+|-',0.2), S298 = (6.38,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-CbH BENSON""", longDesc = u""" """, ) entry( index = 146, label = "Cds-CddCsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cs u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 147, label = "Cds-(Cdd-O2d)CsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([43.83,50.1,55.5,60.05,67.09,72.13,79.55],'J/(mol*K)','+|-',[4,4,4,4,4,4,4]), H298 = (-17.6,'kJ/mol','+|-',3.41), S298 = (169.15,'J/(mol*K)','+|-',4.67), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 148, label = "Cds-(Cdd-Cd)CsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 149, label = "Cds-CddCdsH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 150, label = "Cds-(Cdd-O2d)(Cds-O2d)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {6,D} 4 H u0 {1,S} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 151, label = "Cds-(Cdd-O2d)(Cds-Cd)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 O2d u0 {2,D} 6 C u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([43.67,52.95,59.65,64.67,71.81,76.72,83.92],'J/(mol*K)','+|-',[5.66,5.66,5.66,5.66,5.66,5.66,5.66]), H298 = (-36,'kJ/mol','+|-',4.82), S298 = (152.19,'J/(mol*K)','+|-',6.6), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 152, label = "Cds-(Cdd-O2d)(Cds-Cds)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 O2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 153, label = "Cds-(Cdd-O2d)(Cds-Cdd)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 O2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 154, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)H", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 H u0 {1,S} 6 O2d u0 {3,D} 7 O2d u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.55,12.41,13.82,14.91,16.51,17.62,19.24],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-4.998,'kcal/mol','+|-',0.2), S298 = (39.06,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{CCO/H/CCO} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 155, label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)H", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 H u0 {1,S} 6 O2d u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cd)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 S2d u0 {2,D} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cds)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 S2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 S2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)H", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 H u0 {1,S} 6 S2d u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)H", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 H u0 {1,S} 6 S2d u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 156, label = "Cds-(Cdd-Cd)(Cds-O2d)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {6,D} 4 H u0 {1,S} 5 C u0 {2,D} 6 O2d u0 {3,D} """, thermo = u'Cd-Cd(CO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 157, label = "Cds-(Cdd-Cd)(Cds-Cd)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 C u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 158, label = "Cds-(Cdd-Cd)(Cds-Cds)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 C u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 159, label = "Cds-(Cdd-Cd)(Cds-Cdd)H", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 C u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 160, label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)H", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 H u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cd-Cd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)H", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 H u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 161, label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)H", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 H u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 162, label = "Cds-CddCtH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Ct u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 163, label = "Cds-(Cdd-O2d)CtH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CtH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 164, label = "Cds-(Cdd-Cd)CtH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 165, label = "Cds-CddCbH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cb u0 {1,S} 4 H u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 166, label = "Cds-(Cdd-O2d)CbH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CbH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 167, label = "Cds-(Cdd-Cd)CbH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)C=SH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {6,D} 4 H u0 {1,S} 5 C u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)C=SH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {6,D} 4 H u0 {1,S} 5 S2d u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1182, label = "Cds-CdsC=SH", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 H u0 {1,S} 5 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.41,5.2,5.98,6.68,7.8,8.62,9.84],'cal/(mol*K)'), H298 = (8.87,'kcal/mol'), S298 = (7.87,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1999, label = "Cd-Cd(CO)H", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CO u0 {1,S} {5,D} 3 H u0 {1,S} 4 Cd u0 {1,D} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([18.08,21.17,24.43,27.41,32.22,35.73,40.97],'J/(mol*K)'), H298 = (36.4,'kJ/mol'), S298 = (33.51,'J/(mol*K)'), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 2000, label = "Cd-Cd(CCO)H", group = """ 1 * Cd u0 {2,S} {4,S} {5,D} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 H u0 {1,S} 5 Cd u0 {1,D} 6 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([18.08,21.17,24.43,27.41,32.22,35.73,40.97],'J/(mol*K)'), H298 = (36.4,'kJ/mol'), S298 = (33.51,'J/(mol*K)'), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 168, label = "Cds-CdCO", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 C u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-CdsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 170, label = "Cds-CdsCdsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-Cds(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 171, label = "Cds-Cds(Cds-O2d)O2s", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 CO u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 O2s u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (5.13,'kcal/mol','+|-',0.2), S298 = (-14.6,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-OCO adj BENSON for RADOM c*coh""", longDesc = u""" """, ) entry( index = 172, label = "Cds-Cds(Cds-Cd)O2s", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 O2s u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 173, label = "Cds-Cds(Cds-Cds)O2s", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 O2s u0 {1,S} 5 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (1.5,'kcal/mol','+|-',0.2), S298 = (-14.4,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-OCd jwb need calc""", longDesc = u""" """, ) entry( index = 174, label = "Cds-Cds(Cds-Cdd)O2s", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 O2s u0 {1,S} 5 Cdd u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 175, label = "Cds-Cds(Cds-Cdd-O2d)O2s", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 O2s u0 {1,S} 6 O2d u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 176, label = "Cds-Cds(Cds-Cdd-Cd)O2s", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 O2s u0 {1,S} 6 C u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 177, label = "Cds-CdsCtOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-Cds(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 178, label = "Cds-CdsCbOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (1.5,'kcal/mol','+|-',0.2), S298 = (-14.4,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-OCb jwb need calc""", longDesc = u""" """, ) entry( index = 182, label = "Cds-CddCdsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 183, label = "Cds-(Cdd-O2d)(Cds-O2d)O2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 184, label = "Cds-(Cdd-O2d)(Cds-Cd)O2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 O2d u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 185, label = "Cds-(Cdd-O2d)(Cds-Cds)O2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 O2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 186, label = "Cds-(Cdd-O2d)(Cds-Cdd)O2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 O2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 187, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)O2s", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 O2s u0 {1,S} 6 O2d u0 {3,D} 7 O2d u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.01,12.97,14.17,14.97,15.8,16.26,16.88],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (1.607,'kcal/mol','+|-',0.2), S298 = (17.73,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{CCO/O/CCO} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 188, label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)O2s", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 O2s u0 {1,S} 6 O2d u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 189, label = "Cds-(Cdd-Cd)(Cds-Cd)O2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 C u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 190, label = "Cds-(Cdd-Cd)(Cds-Cds)O2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 C u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 191, label = "Cds-(Cdd-Cd)(Cds-Cdd)O2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 C u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 192, label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)O2s", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 O2s u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-Cds(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 193, label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)O2s", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 O2s u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 194, label = "Cds-CddCtOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 195, label = "Cds-(Cdd-O2d)CtOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 196, label = "Cds-(Cdd-Cd)CtOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 197, label = "Cds-CddCbOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 198, label = "Cds-(Cdd-O2d)CbOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 199, label = "Cds-(Cdd-Cd)CbOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1995, label = "Cd-CdCsOs", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cs u0 {1,S} 3 O2s u0 {1,S} 4 C u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.79,15.86,19.67,22.91,26.55,27.85,28.45],'J/(mol*K)','+|-',[5.1,5.1,5.1,5.1,5.1,5.1,5.1]), H298 = (33,'kJ/mol','+|-',4.34), S298 = (-50.89,'J/(mol*K)','+|-',5.94), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 169, label = "Cds-CdsCsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.59,4.56,5.04,5.3,5.84,6.07,6.16],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (3.03,'kcal/mol','+|-',0.2), S298 = (-12.32,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cd-OCs BOZZELLI-RADOM vin-oh and del (ccoh-ccohc)""", longDesc = u""" """, ) entry( index = 179, label = "Cds-CddCsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 180, label = "Cds-(Cdd-O2d)CsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.91,12.65,13.59,14.22,15,15.48,16.28],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (3.273,'kcal/mol','+|-',0.2), S298 = (18.58,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{CCO/O/C} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 181, label = "Cds-(Cdd-Cd)CsOs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdCS", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 C u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1181, label = "Cds-CdsCsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.23,4.63,4.97,5.29,5.83,6.17,6.53],'cal/(mol*K)'), H298 = (10.63,'kcal/mol'), S298 = (-12.76,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsCdsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cd u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cd)S2s", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 S2s u0 {1,S} 5 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cds)S2s", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 S2s u0 {1,S} 5 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd)S2s", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 S2s u0 {1,S} 5 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd-S2d)S2s", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 S2s u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd-Cd)S2s", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 S2s u0 {1,S} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsCtSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsCbSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CddCsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)CsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CddCdsSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cd u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cd)S2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 S2d u0 {2,D} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cds)S2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 S2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd)S2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 S2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)S2s", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 S2s u0 {1,S} 6 S2d u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)S2s", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 S2s u0 {1,S} 6 S2d u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cd)S2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 C u0 {2,D} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cds)S2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 C u0 {2,D} 6 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd)S2s", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 C u0 {2,D} 6 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)S2s", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 S2s u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)S2s", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 S2s u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CddCtSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CtSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)CtSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CddCbSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CbSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)CbSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)C=SSs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 S2d u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsC=SSs", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 S2s u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 200, label = "Cds-CdCC", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 C u0 {1,S} 4 C u0 {1,S} """, thermo = u'Cds-CdsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 201, label = "Cds-CdsCsCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.1,4.61,4.99,5.26,5.8,6.08,6.36],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (10.34,'kcal/mol','+|-',0.24), S298 = (-12.7,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CsCs BENSON""", longDesc = u""" """, ) entry( index = 202, label = "Cds-CdsCdsCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} """, thermo = u'Cds-Cds(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 204, label = "Cds-Cds(Cds-Cd)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cs u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 205, label = "Cds-Cds(Cds-Cds)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cs u0 {1,S} 5 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (8.88,'kcal/mol','+|-',0.24), S298 = (-14.6,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CdCs BENSON""", longDesc = u""" """, ) entry( index = 206, label = "Cds-Cds(Cds-Cdd)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cs u0 {1,S} 5 Cdd u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd-S2d)Cs", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 Cs u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 208, label = "Cds-Cds(Cds-Cdd-Cd)Cs", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 Cs u0 {1,S} 6 C u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 209, label = "Cds-CdsCdsCds", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 210, label = "Cds-Cds(Cds-O2d)(Cds-O2d)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CO u0 {1,S} {5,D} 3 CO u0 {1,S} {6,D} 4 Cd u0 {1,D} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = u'Cds-CdsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 211, label = "Cds-Cds(Cds-O2d)(Cds-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CO u0 {1,S} {6,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,D} 5 C u0 {3,D} 6 O2d u0 {2,D} """, thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 212, label = "Cds-Cds(Cds-O2d)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CO u0 {1,S} {6,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,D} 5 Cd u0 {3,D} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.7,6.13,6.87,7.1,7.2,7.16,7.06],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (11.6,'kcal/mol','+|-',0.24), S298 = (-16.5,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-COCd from CD/CD2/ jwb est 6/97""", longDesc = u""" AG Vandeputte, added 7 kcal/mol to the following value (see phd M Sabbe) """, ) entry( index = 213, label = "Cds-Cds(Cds-O2d)(Cds-Cdd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CO u0 {1,S} {6,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,D} 5 Cdd u0 {3,D} 6 O2d u0 {2,D} """, thermo = u'Cds-Cds(Cds-O2d)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 214, label = "Cds-Cds(Cds-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cd u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cd u0 {1,D} 6 O2d u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cd-CdCs(CCO)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 215, label = "Cds-Cds(Cds-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cd u0 {1,D} 6 O2d u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 216, label = "Cds-Cds(Cds-Cd)(Cds-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,D} 5 C u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 217, label = "Cds-Cds(Cds-Cds)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,D} 5 Cd u0 {2,D} 6 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([1.9,2.69,3.5,4.28,5.57,6.21,7.37],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (11.6,'kcal/mol','+|-',0.24), S298 = (-15.67,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CdCd Hf=3D est S,Cp mopac nov99""", longDesc = u""" AG Vandeputte, added 7 kcal/mol to the following value (see phd M Sabbe) """, ) entry( index = 218, label = "Cds-Cds(Cds-Cds)(Cds-Cdd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,D} 5 Cd u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 219, label = "Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)", group = """ 1 * Cd u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cd u0 {1,D} 6 Cd u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cd-CdCs(CCO)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cds)(Cds-Cdd-S2d)", group = """ 1 * Cd u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cd u0 {1,D} 6 Cd u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 220, label = "Cds-Cds(Cds-Cds)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cd u0 {1,D} 6 Cd u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 221, label = "Cds-Cds(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,D} 5 Cdd u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 222, label = "Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cd u0 {2,S} {3,S} {6,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {7,D} 5 Cdd u0 {3,D} {8,D} 6 Cd u0 {1,D} 7 O2d u0 {4,D} 8 O2d u0 {5,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 223, label = "Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {6,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {7,D} 5 Cdd u0 {3,D} {8,D} 6 Cd u0 {1,D} 7 O2d u0 {4,D} 8 C u0 {5,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cd u0 {2,S} {3,S} {6,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {7,D} 5 Cdd u0 {3,D} {8,D} 6 Cd u0 {1,D} 7 S2d u0 {4,D} 8 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {6,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {7,D} 5 Cdd u0 {3,D} {8,D} 6 Cd u0 {1,D} 7 S2d u0 {4,D} 8 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 224, label = "Cds-Cds(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {6,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {7,D} 5 Cdd u0 {3,D} {8,D} 6 Cd u0 {1,D} 7 C u0 {4,D} 8 C u0 {5,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 225, label = "Cds-CdsCtCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.5,3.88,4.88,4.18,4.86,5.4,6.01],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (8.11,'kcal/mol','+|-',0.24), S298 = (-13.02,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CtCs RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 1858, label = "Cd-CdCs(CtN3t)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} {5,S} {6,S} 3 Ct u0 {1,S} {7,T} 4 Cs u0 {1,S} 5 R u0 {2,S} 6 R u0 {2,S} 7 N3t u0 {3,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([9.2,10.6,11.7,12.5,13.8,14.7,15.9],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (40.2,'kcal/mol','+|-',1.3), S298 = (17.9,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 226, label = "Cds-CdsCtCds", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-Cds(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 227, label = "Cds-CdsCt(Cds-O2d)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 CO u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Ct u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 228, label = "Cds-CdsCt(Cds-Cd)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Ct u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 229, label = "Cds-Cds(Cds-Cds)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Ct u0 {1,S} 5 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.89,5.26,5.98,6.37,6.67,6.78,6.89],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (7.54,'kcal/mol','+|-',0.24), S298 = (-14.65,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CtCd RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 230, label = "Cds-Cds(Cds-Cdd)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Ct u0 {1,S} 5 Cdd u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cdd-Cd)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 231, label = "Cds-Cds(Cds-Cdd-O2d)Ct", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 Ct u0 {1,S} 6 O2d u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd-S2d)Ct", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 Ct u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 232, label = "Cds-Cds(Cds-Cdd-Cd)Ct", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 Ct u0 {1,S} 6 C u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 233, label = "Cds-CdsCtCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.23,4.59,5.41,5.93,6.48,6.74,7.02],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (8.81,'kcal/mol','+|-',0.24), S298 = (-13.51,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CtCt RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 1837, label = "Cds-Cd(CtN3t)(CtN3t)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Ct u0 {1,S} {5,T} 3 Ct u0 {1,S} {6,T} 4 Cd u0 {1,D} 5 N3t u0 {2,T} 6 N3t u0 {3,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (84.1,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 234, label = "Cds-CdsCbCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (8.64,'kcal/mol','+|-',0.24), S298 = (-14.6,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CbCs BENSON""", longDesc = u""" """, ) entry( index = 235, label = "Cds-CdsCbCds", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cb u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-Cds(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 236, label = "Cds-CdsCb(Cds-O2d)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 CO u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cb u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 237, label = "Cds-Cds(Cds-Cd)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cb u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 238, label = "Cds-Cds(Cds-Cds)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cb u0 {1,S} 5 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.7,6.13,6.87,7.1,7.2,7.16,7.06],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (7.18,'kcal/mol','+|-',0.24), S298 = (-16.5,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CbCd BOZZELLI =3D Cd/Cs/Cb + (Cd/Cs/Cd - Cd/Cs/Cs)""", longDesc = u""" """, ) entry( index = 239, label = "Cds-Cds(Cds-Cdd)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cb u0 {1,S} 5 Cdd u0 {2,D} """, thermo = u'Cds-Cds(Cds-Cdd-Cd)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 240, label = "Cds-Cds(Cds-Cdd-O2d)Cb", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 Cb u0 {1,S} 6 O2d u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-Cds(Cds-Cdd-S2d)Cb", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 Cb u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 241, label = "Cds-Cds(Cds-Cdd-Cd)Cb", group = """ 1 * Cd u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cd u0 {1,D} 5 Cb u0 {1,S} 6 C u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 242, label = "Cds-CdsCbCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2.22,3.14,4.54,4.11,5.06,5.79,6.71],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (6.7,'kcal/mol','+|-',0.24), S298 = (-17.04,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CbCt Hf=3D est S,Cp mopac nov99""", longDesc = u""" """, ) entry( index = 243, label = "Cds-CdsCbCb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.7,6.13,6.87,7.1,7.2,7.16,7.06],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (8,'kcal/mol','+|-',0.24), S298 = (-16.5,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cd-CbCb BOZZELLI =3D Cd/Cs/Cb + (Cd/Cs/Cb - Cd/Cs/Cs)""", longDesc = u""" """, ) entry( index = 244, label = "Cds-CddCsCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 245, label = "Cds-(Cdd-O2d)CsCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([42.55,46.42,50,53.24,58.3,61.71,66.01],'J/(mol*K)','+|-',[4.76,4.76,4.76,4.76,4.76,4.76,4.76]), H298 = (0.5,'kJ/mol','+|-',4.06), S298 = (84.72,'J/(mol*K)','+|-',5.55), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CsCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 246, label = "Cds-(Cdd-Cd)CsCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 247, label = "Cds-CddCdsCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 248, label = "Cds-(Cdd-O2d)(Cds-O2d)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 249, label = "Cds-(Cdd-O2d)(Cds-Cd)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 O2d u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 250, label = "Cds-(Cdd-O2d)(Cds-Cds)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 O2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 251, label = "Cds-(Cdd-O2d)(Cds-Cdd)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 O2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 252, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cs u0 {1,S} 6 O2d u0 {3,D} 7 O2d u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.1,11.24,12.12,12.84,14,14.75,15.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-2.07,'kcal/mol','+|-',0.24), S298 = (19.65,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""{CCO/C/CCO} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 253, label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cs u0 {1,S} 6 O2d u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cd)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 S2d u0 {2,D} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cds)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 S2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 S2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cs u0 {1,S} 6 S2d u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cs u0 {1,S} 6 S2d u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 254, label = "Cds-(Cdd-Cd)(Cds-Cd)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 C u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 255, label = "Cds-(Cdd-Cd)(Cds-Cds)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 C u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 256, label = "Cds-(Cdd-Cd)(Cds-Cdd)Cs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 C u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 257, label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cd-CdCs(CCO)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 258, label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cs", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 259, label = "Cds-CddCdsCds", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 260, label = "Cds-(Cdd-O2d)(Cds-O2d)(Cds-O2d)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {6,D} 4 CO u0 {1,S} {7,D} 5 O2d u0 {2,D} 6 O2d u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 261, label = "Cds-(Cdd-O2d)(Cds-Cd)(Cds-O2d)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {7,D} 5 O2d u0 {2,D} 6 C u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 262, label = "Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {7,D} 5 O2d u0 {2,D} 6 Cd u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 263, label = "Cds-(Cdd-O2d)(Cds-Cdd)(Cds-O2d)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {7,D} 5 O2d u0 {2,D} 6 Cdd u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 264, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-O2d)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 CO u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 O2d u0 {3,D} 7 O2d u0 {4,D} 8 O2d u0 {5,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 265, label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-O2d)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 CO u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 O2d u0 {3,D} 7 O2d u0 {4,D} 8 C u0 {5,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 266, label = "Cds-(Cdd-O2d)(Cds-Cd)(Cds-Cd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 O2d u0 {2,D} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 267, label = "Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 O2d u0 {2,D} 6 Cd u0 {3,D} 7 Cd u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 268, label = "Cds-(Cdd-O2d)(Cds-Cdd)(Cds-Cds)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 O2d u0 {2,D} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 269, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 O2d u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {5,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 270, label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 O2d u0 {3,D} 7 Cd u0 {4,D} 8 C u0 {5,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 271, label = "Cds-(Cdd-O2d)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 O2d u0 {2,D} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 272, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 O2d u0 {4,D} 8 O2d u0 {5,D} 9 O2d u0 {6,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 273, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 O2d u0 {4,D} 8 O2d u0 {5,D} 9 C u0 {6,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 274, label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 O2d u0 {4,D} 8 C u0 {5,D} 9 C u0 {6,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 275, label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-O2d)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {6,D} 4 CO u0 {1,S} {7,D} 5 C u0 {2,D} 6 O2d u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-Cds(Cds-O2d)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 276, label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {7,D} 4 Cd u0 {1,S} {6,D} 5 C u0 {2,D} 6 C u0 {4,D} 7 O2d u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 277, label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cds)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {7,D} 4 Cd u0 {1,S} {6,D} 5 C u0 {2,D} 6 Cd u0 {4,D} 7 O2d u0 {3,D} """, thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 278, label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {7,D} 4 Cd u0 {1,S} {6,D} 5 C u0 {2,D} 6 Cdd u0 {4,D} 7 O2d u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 279, label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 CO u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 C u0 {3,D} 7 O2d u0 {4,D} 8 O2d u0 {5,D} """, thermo = u'Cds-Cds(Cds-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 280, label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 CO u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 C u0 {3,D} 7 O2d u0 {4,D} 8 C u0 {5,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cd)(Cds-Cd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 S2d u0 {2,D} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cds)(Cds-Cds)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 S2d u0 {2,D} 6 Cd u0 {3,D} 7 Cd u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd)(Cds-Cds)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 S2d u0 {2,D} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 S2d u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 S2d u0 {3,D} 7 Cd u0 {4,D} 8 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 S2d u0 {2,D} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 S2d u0 {4,D} 8 S2d u0 {5,D} 9 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 S2d u0 {4,D} 8 S2d u0 {5,D} 9 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 S2d u0 {4,D} 8 C u0 {5,D} 9 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 281, label = "Cds-(Cdd-Cd)(Cds-Cd)(Cds-Cd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 C u0 {2,D} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 282, label = "Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 C u0 {2,D} 6 Cd u0 {3,D} 7 Cd u0 {4,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 283, label = "Cds-(Cdd-Cd)(Cds-Cdd)(Cds-Cds)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 C u0 {2,D} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 284, label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 C u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {5,D} """, thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 C u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 285, label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 C u0 {3,D} 7 Cd u0 {4,D} 8 C u0 {5,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 286, label = "Cds-(Cdd-Cd)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 C u0 {2,D} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 287, label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 C u0 {4,D} 8 O2d u0 {5,D} 9 O2d u0 {6,D} """, thermo = u'Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 288, label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 C u0 {4,D} 8 O2d u0 {5,D} 9 C u0 {6,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 C u0 {4,D} 8 S2d u0 {5,D} 9 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 C u0 {4,D} 8 S2d u0 {5,D} 9 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 289, label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {1,D} {7,D} 5 Cdd u0 {2,D} {8,D} 6 Cdd u0 {3,D} {9,D} 7 C u0 {4,D} 8 C u0 {5,D} 9 C u0 {6,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 290, label = "Cds-CddCtCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 291, label = "Cds-(Cdd-O2d)CtCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CtCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 292, label = "Cds-(Cdd-Cd)CtCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 293, label = "Cds-CddCtCds", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 294, label = "Cds-(Cdd-O2d)(Cds-O2d)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 295, label = "Cds-(Cdd-O2d)(Cds-Cd)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 O2d u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 296, label = "Cds-(Cdd-O2d)(Cds-Cds)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 O2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 297, label = "Cds-(Cdd-O2d)(Cds-Cdd)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 O2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 298, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Ct u0 {1,S} 6 O2d u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 299, label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Ct u0 {1,S} 6 O2d u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cd)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 S2d u0 {2,D} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cds)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 S2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 S2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Ct u0 {1,S} 6 S2d u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Ct u0 {1,S} 6 S2d u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 300, label = "Cds-(Cdd-Cd)(Cds-Cd)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 C u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 301, label = "Cds-(Cdd-Cd)(Cds-Cds)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 C u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 302, label = "Cds-(Cdd-Cd)(Cds-Cdd)Ct", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 C u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 303, label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-Cds(Cds-Cdd-O2d)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 304, label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)Ct", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 305, label = "Cds-CddCtCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 306, label = "Cds-(Cdd-O2d)CtCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CtCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 307, label = "Cds-(Cdd-Cd)CtCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 308, label = "Cds-CddCbCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 309, label = "Cds-(Cdd-O2d)CbCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CbCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 310, label = "Cds-(Cdd-Cd)CbCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 311, label = "Cds-CddCbCds", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cb u0 {1,S} 4 [Cd,CO] u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 312, label = "Cds-(Cdd-O2d)(Cds-O2d)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CO u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 313, label = "Cds-(Cdd-O2d)(Cds-Cd)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 O2d u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 314, label = "Cds-(Cdd-O2d)(Cds-Cds)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 O2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 315, label = "Cds-(Cdd-O2d)(Cds-Cdd)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 O2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 316, label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cb u0 {1,S} 6 O2d u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 317, label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cb u0 {1,S} 6 O2d u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cd)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 S2d u0 {2,D} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cds)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 S2d u0 {2,D} 6 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 S2d u0 {2,D} 6 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cb u0 {1,S} 6 S2d u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cb u0 {1,S} 6 S2d u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 318, label = "Cds-(Cdd-Cd)(Cds-Cd)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 C u0 {2,D} 6 C u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 319, label = "Cds-(Cdd-Cd)(Cds-Cds)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 C u0 {2,D} 6 Cd u0 {3,D} """, thermo = u'Cds-Cds(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 320, label = "Cds-(Cdd-Cd)(Cds-Cdd)Cb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 C u0 {2,D} 6 Cdd u0 {3,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 321, label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cb u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {4,D} """, thermo = u'Cds-Cds(Cds-Cdd-O2d)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cb u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 322, label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cb", group = """ 1 * Cd u0 {2,S} {3,D} {5,S} 2 Cd u0 {1,S} {4,D} 3 Cdd u0 {1,D} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cb u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} """, thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 323, label = "Cds-CddCbCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 324, label = "Cds-(Cdd-O2d)CbCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CbCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 325, label = "Cds-(Cdd-Cd)CbCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 326, label = "Cds-CddCbCb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} """, thermo = u'Cds-(Cdd-Cd)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 327, label = "Cds-(Cdd-O2d)CbCb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 O2d u0 {2,D} """, thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)CbCb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 328, label = "Cds-(Cdd-Cd)CbCb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 C u0 {2,D} """, thermo = u'Cds-CdsCbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsC=SC=S", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CS u0 {1,S} {5,D} 3 CS u0 {1,S} {6,D} 4 Cd u0 {1,D} 5 S2d u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)C=S(Cds-Cd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {7,D} 4 Cd u0 {1,S} {6,D} 5 C u0 {2,D} 6 C u0 {4,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)C=S(Cds-Cds)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {7,D} 4 Cd u0 {1,S} {6,D} 5 C u0 {2,D} 6 Cd u0 {4,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)C=S(Cds-Cdd)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {7,D} 4 Cd u0 {1,S} {6,D} 5 C u0 {2,D} 6 Cdd u0 {4,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)C=S(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 CS u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 C u0 {3,D} 7 S2d u0 {4,D} 8 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)C=S(Cds-Cdd-S2d)", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 CS u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 C u0 {3,D} 7 S2d u0 {4,D} 8 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)C=SCs", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 S2d u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)C=SCt", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 S2d u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)C=SCb", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 S2d u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-Cd)C=SC=S", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {6,D} 4 CS u0 {1,S} {7,D} 5 C u0 {2,D} 6 S2d u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cd)C=S", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {7,D} 5 S2d u0 {2,D} 6 C u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cds)C=S", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {7,D} 5 S2d u0 {2,D} 6 Cd u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd)C=S", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {7,D} 5 S2d u0 {2,D} 6 Cdd u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)C=S", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 CS u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 S2d u0 {3,D} 7 S2d u0 {4,D} 8 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)C=S", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 Cdd u0 {1,D} {6,D} 4 CS u0 {1,S} {7,D} 5 Cdd u0 {2,D} {8,D} 6 S2d u0 {3,D} 7 S2d u0 {4,D} 8 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsCbC=S", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cb u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsCtC=S", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Ct u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1204, label = "Cds-CdsC=SCs", group = """ 1 * Cd u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 Cd u0 {1,D} 4 Cs u0 {1,S} 5 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.44,4.73,4.94,5.14,5.48,5.75,6.24],'cal/(mol*K)'), H298 = (10.34,'kcal/mol'), S298 = (-11.67,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsC=S(Cds-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CS u0 {1,S} {6,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,D} 5 C u0 {3,D} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsC=S(Cds-Cds)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CS u0 {1,S} {6,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,D} 5 Cd u0 {3,D} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsC=S(Cds-Cdd)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CS u0 {1,S} {6,D} 3 Cd u0 {1,S} {5,D} 4 Cd u0 {1,D} 5 Cdd u0 {3,D} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsC=S(Cds-Cdd-Cd)", group = """ 1 * Cd u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cd u0 {1,D} 6 S2d u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-CdsC=S(Cds-Cdd-S2d)", group = """ 1 * Cd u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 Cd u0 {1,D} 6 S2d u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cds-(Cdd-S2d)C=SC=S", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cdd u0 {1,D} {5,D} 3 CS u0 {1,S} {6,D} 4 CS u0 {1,S} {7,D} 5 S2d u0 {2,D} 6 S2d u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1996, label = "Cd-CdCs(CO)", group = """ 1 * Cd u0 {2,S} {3,S} {4,D} 2 CO u0 {1,S} {5,D} 3 Cs u0 {1,S} 4 Cd u0 {1,D} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([15.33,16.82,18.64,20.42,23.2,25,27.1],'J/(mol*K)','+|-',[5.66,5.66,5.66,5.66,5.66,5.66,5.66]), H298 = (39,'kJ/mol','+|-',4.82), S298 = (-51.26,'J/(mol*K)','+|-',6.6), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 2011, label = "Cd-CdCs(CCO)", group = """ 1 * Cd u0 {2,S} {4,S} {5,D} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 Cs u0 {1,S} 5 Cd u0 {1,D} 6 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([22.68,24.05,24.63,25.07,25.64,25.84,25.7],'J/(mol*K)','+|-',[8,8,8,8,8,8,8]), H298 = (41.6,'kJ/mol','+|-',6.82), S298 = (-48.01,'J/(mol*K)','+|-',9.33), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1923, label = "Cds-CNH", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 N u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1860, label = "Cd-CdHN3s", group = """ 1 * Cd u0 {2,D} {5,S} {6,S} 2 Cd u0 {1,D} {3,S} {4,S} 3 R u0 {2,S} 4 R u0 {2,S} 5 H u0 {1,S} 6 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.7,6,7,7.7,8.8,9.5,10.6],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (2.2,'kcal/mol','+|-',1.4), S298 = (7.1,'cal/(mol*K)','+|-',1.3), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1838, label = "Cd-CdH(N5dOdOs)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} {5,S} {6,S} 3 N5dc u0 {1,S} {7,D} {8,S} 4 H u0 {1,S} 5 R u0 {2,S} 6 R u0 {2,S} 7 O2d u0 {3,D} 8 O2s u0 {3,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.7,15.4,17.6,19.3,21.7,23.1,25],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (2,'kcal/mol','+|-',1.3), S298 = (44.3,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1924, label = "Cds-CCN", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 C u0 {1,D} 3 C u0 {1,S} 4 N u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1859, label = "Cd-CdCsN3s", group = """ 1 * Cd u0 {2,D} {5,S} {6,S} 2 Cd u0 {1,D} {3,S} {4,S} 3 R u0 {2,S} 4 R u0 {2,S} 5 Cs u0 {1,S} 6 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.8,5,5.9,6.4,6.9,7.1,7.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (3.5,'kcal/mol','+|-',1.4), S298 = (-14.1,'cal/(mol*K)','+|-',1.3), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1861, label = "Cd-CdCs(N5dOdOs)", group = """ 1 * Cd u0 {2,D} {3,S} {4,S} 2 Cd u0 {1,D} {5,S} {6,S} 3 N5dc u0 {1,S} {7,D} {8,S} 4 Cs u0 {1,S} 5 R u0 {2,S} 6 R u0 {2,S} 7 O2d u0 {3,D} 8 O2s u0 {3,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.1,14.3,16.1,17.5,19.3,20.3,21.4],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (2.3,'kcal/mol','+|-',1.3), S298 = (24,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-SsSs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CH", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 C u0 {1,S} 4 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1184, label = "C=S-CsH", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cs u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.11,9.03,9.88,10.61,11.74,12.55,13.82],'cal/(mol*K)'), H298 = (27.32,'kcal/mol'), S298 = (37.56,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1185, label = "C=S-CdsH", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cd u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.59,9.38,10.81,11.85,13.18,13.95,14.81],'cal/(mol*K)'), H298 = (24.05,'kcal/mol'), S298 = (34.35,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cd)H", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 H u0 {1,S} 5 [Cd,Cdd,CO] u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd)H", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 H u0 {1,S} 5 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-Cd)H", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 H u0 {1,S} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-S2d)H", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 H u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cds)H", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 H u0 {1,S} 5 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1186, label = "C=S-CtH", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Ct u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.46,8.91,10.01,10.83,11.98,12.74,13.87],'cal/(mol*K)'), H298 = (30.83,'kcal/mol'), S298 = (37.16,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1187, label = "C=S-CbH", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cb u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.45,9.84,10.94,11.78,12.97,13.76,14.77],'cal/(mol*K)'), H298 = (24.71,'kcal/mol'), S298 = (34.15,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1188, label = "C=S-C=SH", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 H u0 {1,S} 5 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.79,9.18,10.41,11.42,12.82,13.63,14.54],'cal/(mol*K)'), H298 = (26.96,'kcal/mol'), S298 = (35.65,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CC", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 C u0 {1,S} 4 C u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CbCds", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cb u0 {1,S} 4 Cd u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Cb(Cds-Cd)", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cb u0 {1,S} 5 [Cd,Cdd,CO] u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Cb(Cds-Cds)", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cb u0 {1,S} 5 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Cb(Cds-Cdd)", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cb u0 {1,S} 5 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Cb(Cds-Cdd-S2d)", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 Cb u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Cb(Cds-Cdd-Cd)", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 Cb u0 {1,S} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CtCt", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CbCb", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CdsCds", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cd u0 {1,S} 4 Cd u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cd)(Cds-Cd)", group = """ 1 * CS u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2d u0 {1,D} 5 [Cd,Cdd,CO] u0 {2,D} 6 [Cd,Cdd,CO] u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd)(Cds-Cds)", group = """ 1 * CS u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2d u0 {1,D} 5 Cdd u0 {2,D} 6 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-Cd)(Cds-Cds)", group = """ 1 * CS u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 S2d u0 {1,D} 6 Cd u0 {3,D} 7 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-S2d)(Cds-Cds)", group = """ 1 * CS u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {6,D} 4 Cdd u0 {2,D} {7,D} 5 S2d u0 {1,D} 6 Cd u0 {3,D} 7 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cds)(Cds-Cds)", group = """ 1 * CS u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2d u0 {1,D} 5 Cd u0 {2,D} 6 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd)(Cds-Cdd)", group = """ 1 * CS u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 S2d u0 {1,D} 5 Cdd u0 {2,D} 6 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * CS u0 {2,S} {3,S} {6,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {7,D} 5 Cdd u0 {3,D} {8,D} 6 S2d u0 {1,D} 7 C u0 {4,D} 8 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * CS u0 {2,S} {3,S} {6,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {7,D} 5 Cdd u0 {3,D} {8,D} 6 S2d u0 {1,D} 7 S2d u0 {4,D} 8 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-Cd)(Cds-Cdd-S2d)", group = """ 1 * CS u0 {2,S} {3,S} {6,D} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {7,D} 5 Cdd u0 {3,D} {8,D} 6 S2d u0 {1,D} 7 C u0 {4,D} 8 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CtCds", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Ct u0 {1,S} 4 Cd u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Ct(Cds-Cd)", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Ct u0 {1,S} 5 [Cd,Cdd,CO] u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Ct(Cds-Cds)", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Ct u0 {1,S} 5 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Ct(Cds-Cdd)", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Ct u0 {1,S} 5 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Ct(Cds-Cdd-Cd)", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 Ct u0 {1,S} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-Ct(Cds-Cdd-S2d)", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 Ct u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CbCt", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1190, label = "C=S-CsCs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.7,7.44,8.14,8.72,9.52,9.98,10.51],'cal/(mol*K)'), H298 = (27.2,'kcal/mol'), S298 = (18,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1191, label = "C=S-CdsCs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cd u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.79,9.21,10.13,10.71,11.25,11.42,11.35],'cal/(mol*K)'), H298 = (26.19,'kcal/mol'), S298 = (13.44,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cd)Cs", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cs u0 {1,S} 5 [Cd,Cdd,CO] u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cds)Cs", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cs u0 {1,S} 5 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd)Cs", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cs u0 {1,S} 5 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-S2d)Cs", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 Cs u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-Cd)Cs", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 Cs u0 {1,S} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1192, label = "C=S-CtCs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.87,7.88,8.6,9.13,9.8,10.17,10.59],'cal/(mol*K)'), H298 = (30.12,'kcal/mol'), S298 = (17.46,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1193, label = "C=S-CbCs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.02,9.02,9.75,10.23,10.75,10.96,11.04],'cal/(mol*K)'), H298 = (26.6,'kcal/mol'), S298 = (14.55,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1194, label = "C=S-C=SCs", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cs u0 {1,S} 5 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.93,7.93,8.76,9.37,10.11,10.45,10.71],'cal/(mol*K)'), H298 = (27.48,'kcal/mol'), S298 = (16.58,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CtC=S", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Ct u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cd)C=S", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cd u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd)C=S", group = """ 1 * CS u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {6,D} 4 S2d u0 {1,D} 5 Cdd u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-Cd)C=S", group = """ 1 * CS u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {6,D} 5 S2d u0 {1,D} 6 C u0 {4,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-S2d)C=S", group = """ 1 * CS u0 {2,S} {3,S} {5,D} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {6,D} 5 S2d u0 {1,D} 6 S2d u0 {4,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cds)C=S", group = """ 1 * CS u0 {2,S} {3,S} {4,D} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {6,D} 4 S2d u0 {1,D} 5 Cd u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-C=SC=S", group = """ 1 * CS u0 {2,S} {3,S} {4,D} 2 CS u0 {1,S} {5,D} 3 CS u0 {1,S} {6,D} 4 S2d u0 {1,D} 5 S2d u0 {2,D} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CbC=S", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 Cb u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1183, label = "C=S-HH", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([9.08,10.34,11.51,12.5,14.07,15.25,17.14],'cal/(mol*K)'), H298 = (27.71,'kcal/mol'), S298 = (56.51,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1189, label = "C=S-SsH", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 S2s u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.38,9.78,10.83,11.66,12.86,13.71,14.87],'cal/(mol*K)'), H298 = (21.55,'kcal/mol'), S298 = (34.41,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1205, label = "C=S-CSs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 C u0 {1,S} 4 S2s u0 {1,S} """, thermo = u'C=S-CsSs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CbSs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CdsSs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cd u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cd)S2s", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 S2s u0 {1,S} 5 [Cd,Cdd,CO] u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cds)S2s", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 S2s u0 {1,S} 5 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd)S2s", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 Cd u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 S2s u0 {1,S} 5 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-Cd)S2s", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 S2s u0 {1,S} 6 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-(Cds-Cdd-S2d)S2s", group = """ 1 * CS u0 {2,S} {4,D} {5,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {6,D} 4 S2d u0 {1,D} 5 S2s u0 {1,S} 6 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "C=S-CtSs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1206, label = "C=S-CsSs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.4,8.38,9.16,9.8,10.72,11.25,11.66],'cal/(mol*K)'), H298 = (21.35,'kcal/mol'), S298 = (14.52,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "C=S-C=SSs", group = """ 1 * CS u0 {2,S} {3,D} {4,S} 2 CS u0 {1,S} {5,D} 3 S2d u0 {1,D} 4 S2s u0 {1,S} 5 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1457, label = "CS-OsH", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 O2s u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.48,5.3,6.09,6.82,8.05,8.99,10.37],'cal/(mol*K)'), H298 = (2.85,'kcal/mol'), S298 = (30.14,'cal/(mol*K)'), ), shortDesc = u"""CAC 1d-HR calc""", longDesc = u""" """, ) entry( index = 1458, label = "CS-CsOs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 O2s u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.9,4.17,4.6,5.1,6.08,6.76,7.44],'cal/(mol*K)'), H298 = (-1.32,'kcal/mol'), S298 = (8.62,'cal/(mol*K)'), ), shortDesc = u"""CAC 1d-HR calc""", longDesc = u""" """, ) entry( index = 1459, label = "CS-OsOs", group = """ 1 * CS u0 {2,D} {3,S} {4,S} 2 S2d u0 {1,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.08,3.59,3.9,4.03,3.99,3.75,3.23],'cal/(mol*K)'), H298 = (-22.72,'kcal/mol'), S298 = (2.67,'cal/(mol*K)'), ), shortDesc = u"""CAC CBS-QB3 1Dhr calc""", longDesc = u""" """, ) entry( index = 329, label = "Cs", group = """ 1 * Cs u0 """, thermo = u'Cs-CsCsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1919, label = "Cs-NHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1800, label = "Cs-N3sHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3s u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)'), H298 = (-10.08,'kcal/mol'), S298 = (30.41,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1920, label = "Cs-N3dHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1870, label = "Cs-(N3dCd)HHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} {6,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 [Cd,Cdd] u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6,7.7,9.3,10.7,13.1,14.8,17.7],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (-5.7,'kcal/mol','+|-',1.3), S298 = (30.4,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1801, label = "Cs-(N3dN3d)HHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} {6,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 N3d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6,7.8,9.4,10.8,13.1,14.8,17.6],'cal/(mol*K)','+|-',[0.6,0.6,0.6,0.6,0.6,0.6,0.6]), H298 = (-9,'kcal/mol','+|-',0.8), S298 = (30.2,'cal/(mol*K)','+|-',0.8), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1921, label = "Cs-NCsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1802, label = "Cs-N3sCsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3s u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.25,6.9,8.28,9.39,11.09,12.34,14.8],'cal/(mol*K)'), H298 = (-6.6,'kcal/mol'), S298 = (9.8,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1925, label = "Cs-N3dCHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1805, label = "Cs-(N3dN3d)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 N3d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.3,6.9,8.3,9.4,11.1,12.3,14.2],'cal/(mol*K)','+|-',[0.6,0.6,0.6,0.6,0.6,0.6,0.6]), H298 = (-5.5,'kcal/mol','+|-',0.8), S298 = (9.4,'cal/(mol*K)','+|-',0.8), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1869, label = "Cs-(N3dOd)CHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.8,13.6,15.2,16.7,18.9,20.5,23],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (21.4,'kcal/mol','+|-',1.3), S298 = (44.3,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1871, label = "Cs-(N3dCd)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.3,7.2,8.7,9.8,11.6,12.8,14.7],'cal/(mol*K)','+|-',[1.2,1.2,1.2,1.2,1.2,1.2,1.2]), H298 = (-2.9,'kcal/mol','+|-',1.7), S298 = (8.6,'cal/(mol*K)','+|-',1.6), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1926, label = "Cs-N5dCsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N5dc u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1841, label = "Cs-(N5dOdOs)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N5dc u0 {1,S} {6,D} {7,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2s u0 {2,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.9,15.8,18.3,20.3,23.3,25.4,28.3],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (-14.8,'kcal/mol','+|-',1.3), S298 = (48.9,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1927, label = "Cs-NCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1803, label = "Cs-N3sCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3s u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.67,6.32,7.64,8.39,9.56,10.23,11.905],'cal/(mol*K)'), H298 = (-5.2,'kcal/mol'), S298 = (-11.7,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1928, label = "Cs-N3dCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1868, label = "Cs-(N3dOd)CsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.2,12.7,14,15.1,16.8,17.9,19.5],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (23.4,'kcal/mol','+|-',1.3), S298 = (23.1,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1929, label = "Cs-N5dCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N5dc u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1842, label = "Cs-(N5dOdOs)CsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N5dc u0 {1,S} {6,D} {7,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2s u0 {2,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([13.6,16.1,18.1,19.6,21.8,23.2,25.1],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (-13.9,'kcal/mol','+|-',1.3), S298 = (27.5,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1930, label = "Cs-NCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1804, label = "Cs-N3sCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3s u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.35,6.16,7.31,7.91,8.49,8.5,8.525],'cal/(mol*K)'), H298 = (-3.2,'kcal/mol'), S298 = (-34.1,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1806, label = "Cs-(N3dN3d)CsCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 N3d u0 {1,S} {3,D} 3 N3d u0 {2,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (-3.3,'kcal/mol'), S298 = (-11.7,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1807, label = "Cs-(N3dN3d)CsCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 N3d u0 {1,S} {3,D} 3 N3d u0 {2,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (-1.9,'kcal/mol'), S298 = (-34.7,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1931, label = "Cs-N3dCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1867, label = "Cs-(N3dOd)CsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N3d u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.2,13.3,14,14.5,15.3,15.7,16.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (24.1,'kcal/mol','+|-',1.3), S298 = (1.2,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1932, label = "Cs-N5dCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N5dc u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1843, label = "Cs-(N5dOdOs)CsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N5dc u0 {1,S} {6,D} {7,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} 7 O2s u0 {2,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([13.5,16,17.8,18.9,20.3,21.1,21.9],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (-12.7,'kcal/mol','+|-',1.3), S298 = (5.2,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1933, label = "Cs-NNCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N u0 {1,S} 3 N u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1934, label = "Cs-N5dN5dCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N5dc u0 {1,S} 3 N5dc u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1933, label = "Cs-NNCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N u0 {1,S} 3 N u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1846, label = "Cs-(N5dOdOs)(N5dOdOs)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 N5dc u0 {1,S} {6,D} {7,S} 3 N5dc u0 {1,S} {8,D} {9,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2s u0 {2,S} 8 O2d u0 {3,D} 9 O2s u0 {3,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (-14.9,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 330, label = "Cs-HHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 H u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.43,9.84,11.14,12.41,15,17.25,20.63],'cal/(mol*K)','+|-',[0.06,0.06,0.06,0.06,0.06,0.06,0.06]), H298 = (-17.9,'kcal/mol','+|-',0.1), S298 = (49.41,'cal/(mol*K)','+|-',0.05), ), shortDesc = u"""CHEMKIN DATABASE S(group) = S(CH4) + Rln(12)""", longDesc = u""" """, ) entry( index = 331, label = "Cs-CHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CsHHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 332, label = "Cs-CsHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]), H298 = (-10.2,'kcal/mol','+|-',0.12), S298 = (30.41,'cal/(mol*K)','+|-',0.08), ), shortDesc = u"""Cs-CsHHH BENSON""", longDesc = u""" """, ) entry( index = 333, label = "Cs-CdsHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)HHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 334, label = "Cs-(Cds-O2d)HHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([25.31,32.07,38.44,44.06,53.36,60.63,72.47],'J/(mol*K)'), H298 = (-42.9,'kJ/mol'), S298 = (127.12,'J/(mol*K)'), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 335, label = "Cs-(Cds-Cd)HHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)HHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 336, label = "Cs-(Cds-Cds)HHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)','+|-',[0.04,0.04,0.04,0.04,0.04,0.04,0.04]), H298 = (-10.2,'kcal/mol','+|-',0.08), S298 = (30.41,'cal/(mol*K)','+|-',0.04), ), shortDesc = u"""Cs-CdHHH BENSON (Assigned Cs-CsHHH)""", longDesc = u""" """, ) entry( index = 337, label = "Cs-(Cds-Cdd)HHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)HHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 338, label = "Cs-(Cds-Cdd-O2d)HHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([25.31,32.07,38.44,44.06,53.36,60.63,72.47],'J/(mol*K)'), H298 = (-42.9,'kJ/mol'), S298 = (127.12,'J/(mol*K)'), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)HHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 339, label = "Cs-(Cds-Cdd-Cd)HHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)HHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1862, label = "Cs-(CdN3d)HHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 N3d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.2,7.8,9.4,10.8,13,14.8,17.6],'cal/(mol*K)'), H298 = (-10.2,'kcal/mol'), S298 = (30.4,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 340, label = "Cs-CtHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]), H298 = (-10.2,'kcal/mol','+|-',0.15), S298 = (30.41,'cal/(mol*K)','+|-',0.08), ), shortDesc = u"""Cs-CtHHH BENSON (Assigned Cs-CsHHH)""", longDesc = u""" """, ) entry( index = 1863, label = "Cs-(CtN3t)HHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} {6,T} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 N3t u0 {2,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.5,14.6,16.6,18.3,21.1,23.4,26.9],'cal/(mol*K)','+|-',[1.3,1.3,1.3,1.3,1.3,1.3,1.3]), H298 = (17.7,'kcal/mol','+|-',1.9), S298 = (60.2,'cal/(mol*K)','+|-',1.7), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 341, label = "Cs-CbHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-10.2,'kcal/mol','+|-',0.18), S298 = (30.41,'cal/(mol*K)','+|-',0.14), ), shortDesc = u"""Cs-CbHHH BENSON (Assigned Cs-CsHHH)""", longDesc = u""" """, ) entry( index = 1176, label = "Cs-C=SHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.96,7.6,9.13,10.49,12.72,14.46,17.28],'cal/(mol*K)'), H298 = (-10.25,'kcal/mol'), S298 = (30.4,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 342, label = "Cs-OsHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 O2s u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([25.31,32.07,38.44,44.06,53.36,60.63,72.47],'J/(mol*K)'), H298 = (-42.9,'kJ/mol'), S298 = (127.12,'J/(mol*K)'), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 343, label = "Cs-OsOsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 O2s u0 {1,S} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([27.35,43.68,53.55,58.15,60.86,61.66,63.53],'J/(mol*K)','+|-',[5.77,5.77,5.77,5.77,5.77,5.77,5.77]), H298 = (-67.5,'kJ/mol','+|-',4.92), S298 = (17.89,'J/(mol*K)','+|-',6.74), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 344, label = "Cs-OsOsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 O2s u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.54,6,7.17,8.05,9.31,10.05,10.05],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-21.23,'kcal/mol','+|-',0.2), S298 = (-12.07,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cs-OOOH BOZZELLI del C/C2/O - C/C3/O, series !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1451, label = "Cs-OsSsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 O2s u0 {1,S} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.48,9.54,11,11.91,12.85,13.54,14.93],'cal/(mol*K)'), H298 = (-11.58,'kcal/mol'), S298 = (4.58,'cal/(mol*K)'), ), shortDesc = u"""CAC CBS-QB3 1DHR CAC""", longDesc = u""" """, ) entry( index = 1464, label = "Cs-OsOsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 O2s u0 {1,S} 3 O2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.36,8.72,10.13,10.88,11.56,11.91,12.53],'cal/(mol*K)'), H298 = (-19.72,'kcal/mol'), S298 = (-13.26,'cal/(mol*K)'), ), shortDesc = u"""CAC calc 1D-HR""", longDesc = u""" """, ) entry( index = 1162, label = "Cs-SsHHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 S2s u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.96,7.6,9.13,10.49,12.72,14.46,17.28],'cal/(mol*K)'), H298 = (-10.25,'kcal/mol'), S298 = (30.4,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1167, label = "Cs-SsSsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 S2s u0 {1,S} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.37,9.7,10.52,11.13,12.16,13.01,14.43],'cal/(mol*K)'), H298 = (-6.21,'kcal/mol'), S298 = (6.14,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1201, label = "Cs-SsSsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 S2s u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.84,9.14,10.24,10.73,11.12,11.33,11.57],'cal/(mol*K)'), H298 = (-2.78,'kcal/mol'), S298 = (-15.38,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 345, label = "Cs-CCHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CsCsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 346, label = "Cs-CsCsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.5,6.95,8.25,9.35,11.07,12.34,14.25],'cal/(mol*K)','+|-',[0.04,0.04,0.04,0.04,0.04,0.04,0.04]), H298 = (-4.93,'kcal/mol','+|-',0.05), S298 = (9.42,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CsCsHH BENSON""", longDesc = u""" """, ) entry( index = 347, label = "Cs-CdsCsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 348, label = "Cs-(Cds-O2d)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([26.91,30.8,34.98,38.91,45.56,50.73,58.93],'J/(mol*K)','+|-',[1.53,1.53,1.53,1.53,1.53,1.53,1.53]), H298 = (-21.5,'kJ/mol','+|-',1.3), S298 = (40.32,'J/(mol*K)','+|-',1.78), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 349, label = "Cs-(Cds-Cd)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 350, label = "Cs-(Cds-Cds)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.12,6.86,8.32,9.49,11.22,12.48,14.36],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-4.76,'kcal/mol','+|-',0.16), S298 = (9.8,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cs-CdCsHH BENSON""", longDesc = u""" """, ) entry( index = 351, label = "Cs-(Cds-Cdd)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 352, label = "Cs-(Cds-Cdd-O2d)CsHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.35,6.83,8.25,9.45,11.19,12.46,14.34],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-5.723,'kcal/mol','+|-',0.16), S298 = (9.37,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{C/C/H2/CCO} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CsHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 353, label = "Cs-(Cds-Cdd-Cd)CsHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1864, label = "Cs-(CdN3d)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} {7,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 N3d u0 {2,D} 7 R u0 {2,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.5,6.9,8.1,9.2,10.9,12.2,14.1],'cal/(mol*K)','+|-',[1.2,1.2,1.2,1.2,1.2,1.2,1.2]), H298 = (-5.1,'kcal/mol','+|-',1.7), S298 = (10.1,'cal/(mol*K)','+|-',1.6), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 354, label = "Cs-CdsCdsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 355, label = "Cs-(Cds-O2d)(Cds-O2d)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([27.77,32.81,37.1,40.67,46.39,50.85,58.25],'J/(mol*K)','+|-',[4.19,4.19,4.19,4.19,4.19,4.19,4.19]), H298 = (-10,'kJ/mol','+|-',3.57), S298 = (40.1,'J/(mol*K)','+|-',4.88), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 356, label = "Cs-(Cds-O2d)(Cds-Cd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([24.94,31.41,36.47,40.49,46.72,51.49,59.29],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]), H298 = (-16.9,'kJ/mol','+|-',2.85), S298 = (40.18,'J/(mol*K)','+|-',3.9), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 357, label = "Cs-(Cds-O2d)(Cds-Cds)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.75,7.11,8.92,10.32,12.16,13.61,13.61],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-3.8,'kcal/mol','+|-',0.16), S298 = (6.31,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cs-COCdHH BENSON Hf, Mopac =3D S,Cp nov99 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 358, label = "Cs-(Cds-O2d)(Cds-Cdd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 359, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)HH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 H u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 360, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 H u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 361, label = "Cs-(Cds-Cd)(Cds-Cd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 362, label = "Cs-(Cds-Cds)(Cds-Cds)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.7,6.8,8.4,9.6,11.3,12.6,14.4],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-4.29,'kcal/mol','+|-',0.16), S298 = (10.2,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cs-CdCdHH BENSON""", longDesc = u""" """, ) entry( index = 363, label = "Cs-(Cds-Cdd)(Cds-Cds)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)HH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 H u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 365, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)HH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 H u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 366, label = "Cs-(Cds-Cdd)(Cds-Cdd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 367, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)HH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 H u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.68,8.28,9.58,10.61,12.04,13.13,14.87],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-5.301,'kcal/mol','+|-',0.16), S298 = (7.18,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{C/H2/CCO2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 368, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 H u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-Cd(CCO)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)HH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 H u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 H u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 369, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 H u0 {1,S} 7 H u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 2010, label = "Cs-Cd(CCO)HH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 H u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([25.85,31.99,37.06,41.14,47.42,52.15,59.73],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]), H298 = (-22.2,'kJ/mol','+|-',5.9), S298 = (37.92,'J/(mol*K)','+|-',8.08), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 370, label = "Cs-CtCsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.95,6.56,7.93,9.08,10.86,12.19,14.2],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]), H298 = (-4.73,'kcal/mol','+|-',0.28), S298 = (10.3,'cal/(mol*K)','+|-',0.07), ), shortDesc = u"""Cs-CtCsHH BENSON""", longDesc = u""" """, ) entry( index = 1832, label = "Cs-(CtN3t)CsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} {6,T} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 N3t u0 {2,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.3,13.5,15.3,16.8,19.2,20.9,23.5],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (22.9,'kcal/mol','+|-',1.3), S298 = (39.8,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 371, label = "Cs-CtCdsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 372, label = "Cs-(Cds-O2d)CtHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.85,6.22,8.01,9.43,11.29,12.76,12.76],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]), H298 = (-5.4,'kcal/mol','+|-',0.28), S298 = (7.68,'cal/(mol*K)','+|-',0.07), ), shortDesc = u"""Cs-COCtHH BENSON Hf, Mopac =3D S,Cp nov99 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 373, label = "Cs-(Cds-Cd)CtHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CtHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 374, label = "Cs-(Cds-Cds)CtHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.4,6.33,7.9,9.16,10.93,12.29,13.43],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]), H298 = (-3.49,'kcal/mol','+|-',0.28), S298 = (9.31,'cal/(mol*K)','+|-',0.07), ), shortDesc = u"""Cs-CtCdHH RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 375, label = "Cs-(Cds-Cdd)CtHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 376, label = "Cs-(Cds-Cdd-O2d)CtHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-Cd(CCO)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 377, label = "Cs-(Cds-Cdd-Cd)CtHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 378, label = "Cs-CtCtHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4,6.07,7.71,9.03,10.88,12.3,12.48],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]), H298 = (-0.82,'kcal/mol','+|-',0.28), S298 = (10.04,'cal/(mol*K)','+|-',0.07), ), shortDesc = u"""Cs-CtCtHH RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 379, label = "Cs-CbCsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.84,7.61,8.98,10.01,11.49,12.54,13.76],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-4.86,'kcal/mol','+|-',0.2), S298 = (9.34,'cal/(mol*K)','+|-',0.19), ), shortDesc = u"""Cs-CbCsHH BENSON""", longDesc = u""" """, ) entry( index = 380, label = "Cs-CbCdsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 381, label = "Cs-(Cds-O2d)CbHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.38,7.59,9.25,10.51,12.19,13.52,13.52],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-5.4,'kcal/mol','+|-',0.2), S298 = (5.89,'cal/(mol*K)','+|-',0.19), ), shortDesc = u"""Cs-COCbHH BENSON Hf, Mopac =3D S,Cp nov99 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 382, label = "Cs-(Cds-Cd)CbHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 383, label = "Cs-(Cds-Cds)CbHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.51,6.76,8.61,10.01,11.97,13.4,15.47],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]), H298 = (-4.29,'kcal/mol','+|-',0.2), S298 = (2,'cal/(mol*K)','+|-',0.19), ), shortDesc = u"""Cs-CbCdHH Hf=Stein S,Cp=3D mopac nov99""", longDesc = u""" """, ) entry( index = 384, label = "Cs-(Cds-Cdd)CbHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 385, label = "Cs-(Cds-Cdd-O2d)CbHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-Cd(CCO)HH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 386, label = "Cs-(Cds-Cdd-Cd)CbHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 387, label = "Cs-CbCtHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.28,6.43,8.16,9.5,11.36,12.74,13.7],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]), H298 = (-4.29,'kcal/mol','+|-',0.28), S298 = (9.84,'cal/(mol*K)','+|-',0.07), ), shortDesc = u"""Cs-CbCtHH Hf=Stein S,Cp=3D mopac nov99""", longDesc = u""" """, ) entry( index = 388, label = "Cs-CbCbHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.67,7.7,9.31,10.52,12.21,13.47,15.11],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]), H298 = (-4.29,'kcal/mol','+|-',0.2), S298 = (8.07,'cal/(mol*K)','+|-',0.19), ), shortDesc = u"""Cs-CbCbHH Hf=3Dbsn/Cs/Cd2/H2 S,Cp=3D mopac nov99""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1177, label = "Cs-C=SCsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.23,6.82,8.16,9.27,10.96,12.2,14.13],'cal/(mol*K)'), H298 = (-4.89,'kcal/mol'), S298 = (9.83,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)HH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 H u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)HH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 H u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)HH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 H u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 389, label = "Cs-CCCH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 C u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CsCsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 390, label = "Cs-CsCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.54,6,7.17,8.05,9.31,10.05,11.17],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (-1.9,'kcal/mol','+|-',0.15), S298 = (-12.07,'cal/(mol*K)','+|-',0.07), ), shortDesc = u"""Cs-CsCsCsH BENSON""", longDesc = u""" """, ) entry( index = 391, label = "Cs-CdsCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 392, label = "Cs-(Cds-O2d)CsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([23.68,27.86,31.26,34,38.07,41,45.46],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]), H298 = (-5.4,'kJ/mol','+|-',2.85), S298 = (-47.41,'J/(mol*K)','+|-',3.9), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 393, label = "Cs-(Cds-Cd)CsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 394, label = "Cs-(Cds-Cds)CsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.16,5.91,7.34,8.19,9.46,10.19,11.28],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-1.48,'kcal/mol','+|-',0.27), S298 = (-11.69,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CdCsCsH BENSON""", longDesc = u""" """, ) entry( index = 395, label = "Cs-(Cds-Cdd)CsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 396, label = "Cs-(Cds-Cdd-O2d)CsCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([21.23,27.55,32.36,35.85,40.37,43.16,46.94],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]), H298 = (-11.1,'kJ/mol','+|-',5.9), S298 = (-47.59,'J/(mol*K)','+|-',8.08), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CsCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 397, label = "Cs-(Cds-Cdd-Cd)CsCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.16,5.91,7.34,8.19,9.46,10.19,11.28],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-1.48,'kcal/mol','+|-',0.27), S298 = (-11.69,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CdCsCsH BENSON""", longDesc = u""" """, ) entry( index = 1865, label = "Cs-(CdN3d)CsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} {7,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 N3d u0 {2,D} 7 R u0 {2,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5,6.5,7.5,8.2,9.3,9.9,10.9],'cal/(mol*K)','+|-',[1.2,1.2,1.2,1.2,1.2,1.2,1.2]), H298 = (-1.6,'kcal/mol','+|-',1.7), S298 = (-11.2,'cal/(mol*K)','+|-',1.6), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 398, label = "Cs-CtCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,5.61,6.85,7.78,9.1,9.9,11.12],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-1.72,'kcal/mol','+|-',0.27), S298 = (-11.19,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CtCsCsH BENSON""", longDesc = u""" """, ) entry( index = 1833, label = "Cs-(CtN3t)CsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} {6,T} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 N3t u0 {2,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11,12.7,14.1,15.4,17.3,18.6,21.85],'cal/(mol*K)'), H298 = (25.8,'kcal/mol'), S298 = (19.8,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 399, label = "Cs-CbCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.88,6.66,7.9,8.75,9.73,10.25,10.68],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-0.98,'kcal/mol','+|-',0.27), S298 = (-12.15,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CbCsCsH BENSON""", longDesc = u""" """, ) entry( index = 400, label = "Cs-CdsCdsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 401, label = "Cs-(Cds-O2d)(Cds-O2d)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-CsCsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 402, label = "Cs-(Cds-O2d)(Cds-Cd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([29.32,32.99,35.49,37.28,39.75,41.6,44.96],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]), H298 = (-2.2,'kJ/mol','+|-',2.85), S298 = (-50.47,'J/(mol*K)','+|-',3.9), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 403, label = "Cs-(Cds-O2d)(Cds-Cds)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 404, label = "Cs-(Cds-O2d)(Cds-Cdd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 405, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 406, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 407, label = "Cs-(Cds-Cd)(Cds-Cd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 408, label = "Cs-(Cds-Cds)(Cds-Cds)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.28,6.54,7.67,8.48,9.45,10.18,11.24],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-1.1,'kcal/mol','+|-',0.27), S298 = (-13.03,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CdCdCsH RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 409, label = "Cs-(Cds-Cdd)(Cds-Cds)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 411, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 412, label = "Cs-(Cds-Cdd)(Cds-Cdd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 413, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([21.19,28,33.91,38.75,46.07,51.36,59.45],'J/(mol*K)','+|-',[3.46,3.46,3.46,3.46,3.46,3.46,3.46]), H298 = (-21.1,'kJ/mol','+|-',2.95), S298 = (40.95,'J/(mol*K)','+|-',4.04), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 414, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-CsCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 415, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 H u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 2008, label = "Cs-CsCd(CCO)H", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([24.45,31.59,36.01,38.8,42.13,44.21,47.25],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]), H298 = (-10.4,'kJ/mol','+|-',5.9), S298 = (-54.03,'J/(mol*K)','+|-',8.08), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 416, label = "Cs-CtCdsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 417, label = "Cs-(Cds-O2d)CtCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 418, label = "Cs-(Cds-Cd)CtCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CtCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 419, label = "Cs-(Cds-Cds)CtCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.55,7.21,8.39,9.17,10,10.61,10.51],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-6.9,'kcal/mol','+|-',0.27), S298 = (-13.48,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CtCdCsH RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 420, label = "Cs-(Cds-Cdd)CtCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 421, label = "Cs-(Cds-Cdd-O2d)CtCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-CsCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 422, label = "Cs-(Cds-Cdd-Cd)CtCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 423, label = "Cs-CbCdsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 424, label = "Cs-(Cds-O2d)CbCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 425, label = "Cs-(Cds-Cd)CbCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 426, label = "Cs-(Cds-Cds)CbCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.5,6.57,8.07,8.89,9.88,10.39,10.79],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-1.56,'kcal/mol','+|-',0.27), S298 = (-11.77,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CbCdCsH BOZZELLI =3D Cs/Cs2/Cd/H + (Cs/Cs2/Cb/H - Cs/Cs3/H)""", longDesc = u""" """, ) entry( index = 427, label = "Cs-(Cds-Cdd)CbCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 428, label = "Cs-(Cds-Cdd-O2d)CbCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-CsCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 429, label = "Cs-(Cds-Cdd-Cd)CbCsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 430, label = "Cs-CtCtCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.27,5.32,6.9,8.03,9.33,10.21,9.38],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (1.72,'kcal/mol','+|-',0.27), S298 = (-11.61,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CtCtCsH RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 431, label = "Cs-CbCtCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.33,6.27,7.58,8.48,9.52,10.1,10.63],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-1.55,'kcal/mol','+|-',0.27), S298 = (-11.65,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CbCtCsH BOZZELLI =3D Cs/Cs2/Cb/H + (Cs/Cs2/Ct/H - Cs/Cs3/H)""", longDesc = u""" """, ) entry( index = 432, label = "Cs-CbCbCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.22,7.32,8.63,8.45,10.15,10.45,10.89],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-1.06,'kcal/mol','+|-',0.27), S298 = (-12.23,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CbCbCsCs BOZZELLI =3D Cs/Cs2/Cb/H + (Cs/Cs2/Cb/H - Cs/Cs3/H)""", longDesc = u""" """, ) entry( index = 433, label = "Cs-CdsCdsCdsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 434, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-CsCsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 435, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 H u0 {1,S} 6 C u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 436, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 H u0 {1,S} 6 Cd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 437, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 H u0 {1,S} 6 Cdd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 438, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 439, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 440, label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 H u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([29.26,34.41,37.4,39.22,41.43,43.04,46.12],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]), H298 = (2.9,'kJ/mol','+|-',2.85), S298 = (-53.2,'J/(mol*K)','+|-',3.9), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 441, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsCsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 442, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 443, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 444, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 445, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 446, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 447, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 448, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 449, label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 H u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 450, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.51,5.96,7.13,7.98,9.06,9.9,11.23],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (0.41,'kcal/mol','+|-',0.27), S298 = (-11.82,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CdCdCdH RAMAN & GREEN JPC 2002""", longDesc = u""" """, ) entry( index = 451, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 453, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 454, label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 455, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 456, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-CdCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 457, label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 Cd u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 458, label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 459, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 H u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 460, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 H u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 461, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 H u0 {1,S} 9 O2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-CdCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 H u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 H u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 H u0 {1,S} 9 S2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 462, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 H u0 {1,S} 9 C u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 2009, label = "Cs-CdCd(CCO)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 O2d u0 {5,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([27.62,35.4,39.24,41.25,43.4,44.87,47.43],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]), H298 = (-6.8,'kJ/mol','+|-',5.9), S298 = (-55.37,'J/(mol*K)','+|-',8.08), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 463, label = "Cs-CtCdsCdsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 464, label = "Cs-(Cds-O2d)(Cds-O2d)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 465, label = "Cs-(Cds-O2d)(Cds-Cd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 466, label = "Cs-(Cds-O2d)(Cds-Cds)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 467, label = "Cs-(Cds-O2d)(Cds-Cdd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 468, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 469, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 470, label = "Cs-(Cds-Cd)(Cds-Cd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 471, label = "Cs-(Cds-Cds)(Cds-Cds)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.68,7.85,8.62,9.16,9.81,10.42,10.49],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (1.88,'kcal/mol','+|-',0.27), S298 = (-13.75,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CtCdCdH RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 472, label = "Cs-(Cds-Cdd)(Cds-Cds)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 473, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-CdCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 474, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 475, label = "Cs-(Cds-Cdd)(Cds-Cdd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 476, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 477, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 478, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 H u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 479, label = "Cs-CbCdsCdsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 480, label = "Cs-(Cds-O2d)(Cds-O2d)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 481, label = "Cs-(Cds-O2d)(Cds-Cd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 482, label = "Cs-(Cds-O2d)(Cds-Cds)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 483, label = "Cs-(Cds-O2d)(Cds-Cdd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 484, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 485, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 486, label = "Cs-(Cds-Cd)(Cds-Cd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 487, label = "Cs-(Cds-Cds)(Cds-Cds)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.12,6.51,8.24,9,10.03,10.53,10.89],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-1.39,'kcal/mol','+|-',0.27), S298 = (-11.39,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CbCdCdH BOZZELLI =3D Cs/Cs/Cd2/H + (Cs/Cs2/Cb/H - Cs/Cs3/H)""", longDesc = u""" """, ) entry( index = 488, label = "Cs-(Cds-Cdd)(Cds-Cds)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 489, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-CdCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 490, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 491, label = "Cs-(Cds-Cdd)(Cds-Cdd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 492, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 493, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 494, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 H u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 495, label = "Cs-CtCtCdsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CtCt(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 496, label = "Cs-CtCt(Cds-O2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 497, label = "Cs-CtCt(Cds-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-CtCt(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 498, label = "Cs-CtCt(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.58,5.68,7.11,8.12,9.27,10.13,9.44],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (4.73,'kcal/mol','+|-',0.27), S298 = (-11.46,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CtCtCdH RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 499, label = "Cs-CtCt(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-CtCt(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 500, label = "Cs-CtCt(Cds-Cdd-O2d)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-CdCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCt(Cds-Cdd-S2d)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 501, label = "Cs-CtCt(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-CtCt(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 502, label = "Cs-CbCtCdsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CbCt(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 503, label = "Cs-CbCt(Cds-O2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 504, label = "Cs-CbCt(Cds-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-CbCt(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 505, label = "Cs-CbCt(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 506, label = "Cs-CbCt(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-CbCt(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 507, label = "Cs-CbCt(Cds-Cdd-O2d)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCt(Cds-Cdd-S2d)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 508, label = "Cs-CbCt(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-CbCt(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 509, label = "Cs-CbCbCdsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CbCb(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 510, label = "Cs-CbCb(Cds-O2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCb(Cds-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 511, label = "Cs-CbCb(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 512, label = "Cs-CbCb(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-CbCb(Cds-Cdd-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 513, label = "Cs-CbCb(Cds-Cdd-O2d)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-CdCd(CCO)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCb(Cds-Cdd-S2d)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 514, label = "Cs-CbCb(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-CbCb(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 515, label = "Cs-CtCtCtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.03,5.27,6.78,7.88,9.14,10.08,8.47],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (10.11,'kcal/mol','+|-',0.27), S298 = (-10.46,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CtCtCtH RAMAN & GREEN JPCA 2002, 106, 11141-11149""", longDesc = u""" """, ) entry( index = 516, label = "Cs-CbCtCtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CtCt(Cds-Cds)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 517, label = "Cs-CbCbCtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 518, label = "Cs-CbCbCbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.56,7.98,9.36,10.15,10.57,10.65,9.7],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-0.34,'kcal/mol','+|-',0.27), S298 = (-12.31,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CbCbCbH BOZZELLI =3D Cs/Cs/Cb2/H + (Cs/Cs2/Cb/H - Cs/Cs3/H)""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)(Cds-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 H u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCtH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CbH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCtC=SH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=SH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 H u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1178, label = "Cs-C=SCsCsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.78,6.25,7.44,8.35,9.57,10.31,11.2],'cal/(mol*K)'), H298 = (-0.78,'kcal/mol'), S298 = (-11.46,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCtC=SH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCbC=SH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 H u0 {1,S} 6 C u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cds)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 H u0 {1,S} 6 Cd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 H u0 {1,S} 6 Cdd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-S2d)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-Cd)H", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 519, label = "Cs-CCCC", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 C u0 {1,S} 5 C u0 {1,S} """, thermo = u'Cs-CsCsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 520, label = "Cs-CsCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.37,6.13,7.36,8.12,8.77,8.76,8.12],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (0.5,'kcal/mol','+|-',0.27), S298 = (-35.1,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CsCsCsCs BENSON""", longDesc = u""" """, ) entry( index = 521, label = "Cs-CdsCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 522, label = "Cs-(Cds-O2d)CsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([22.68,27.48,30.12,31.51,32.36,32.39,32.42],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]), H298 = (4.6,'kJ/mol','+|-',2.85), S298 = (-140.94,'J/(mol*K)','+|-',3.9), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 523, label = "Cs-(Cds-Cd)CsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 524, label = "Cs-(Cds-Cds)CsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,6.04,7.43,8.26,8.92,8.96,8.23],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (1.68,'kcal/mol','+|-',0.27), S298 = (-34.72,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CdCsCsCs BENSON""", longDesc = u""" """, ) entry( index = 525, label = "Cs-(Cds-Cdd)CsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 526, label = "Cs-(Cds-Cdd-O2d)CsCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([20.63,27.65,31.98,34.41,36.16,36.25,35.2],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]), H298 = (-4.5,'kJ/mol','+|-',5.9), S298 = (-144.08,'J/(mol*K)','+|-',8.08), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CsCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 527, label = "Cs-(Cds-Cdd-Cd)CsCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1866, label = "Cs-(CdN3d)CsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} {7,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 N3d u0 {2,D} 7 R u0 {2,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.3,6.6,7.3,7.5,7.8,7.8,7.7],'cal/(mol*K)','+|-',[1.2,1.2,1.2,1.2,1.2,1.2,1.2]), H298 = (0.6,'kcal/mol','+|-',1.7), S298 = (-33.5,'cal/(mol*K)','+|-',1.6), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 528, label = "Cs-CtCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.37,6.79,8.09,8.78,9.19,8.96,7.63],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (2.81,'kcal/mol','+|-',0.27), S298 = (-35.18,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""Cs-CtCsCsCs BENSON""", longDesc = u""" """, ) entry( index = 1834, label = "Cs-(CtN3t)CsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} {6,T} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 N3t u0 {2,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.4,13.4,14.6,15.3,16.3,16.7,17.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (28.3,'kcal/mol','+|-',1.3), S298 = (-3,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 529, label = "Cs-CbCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.37,6.79,8.09,8.78,9.19,8.96,7.63],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (2.81,'kcal/mol','+|-',0.26), S298 = (-35.18,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCsCsCs BENSON""", longDesc = u""" """, ) entry( index = 530, label = "Cs-CdsCdsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 531, label = "Cs-(Cds-O2d)(Cds-O2d)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([33.76,33.42,32.6,31.91,31.01,30.55,30.35],'J/(mol*K)','+|-',[5.08,5.08,5.08,5.08,5.08,5.08,5.08]), H298 = (14.9,'kJ/mol','+|-',4.33), S298 = (-146.69,'J/(mol*K)','+|-',5.92), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 532, label = "Cs-(Cds-O2d)(Cds-Cd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([26.01,30.13,32.44,33.51,33.75,33.26,32.55],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]), H298 = (9.8,'kJ/mol','+|-',2.85), S298 = (-146.74,'J/(mol*K)','+|-',3.9), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 533, label = "Cs-(Cds-O2d)(Cds-Cds)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 534, label = "Cs-(Cds-O2d)(Cds-Cdd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 535, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 536, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 537, label = "Cs-(Cds-Cd)(Cds-Cd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 538, label = "Cs-(Cds-Cds)(Cds-Cds)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,6.04,7.43,8.26,8.92,8.96,8.23],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (1.68,'kcal/mol','+|-',0.26), S298 = (-34.72,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CdCdCsCs BENSON""", longDesc = u""" """, ) entry( index = 539, label = "Cs-(Cds-Cdd)(Cds-Cds)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 541, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 542, label = "Cs-(Cds-Cdd)(Cds-Cdd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 543, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.73,8.1,9.02,9.53,9.66,9.52,8.93],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (-2.987,'kcal/mol','+|-',0.26), S298 = (-36.46,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""{C/C2/CCO2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 544, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-CsCsCd(CCO)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 545, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 Cs u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 2007, label = "Cs-CsCsCd(CCO)", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([25.48,31.89,35.19,36.68,37.19,36.66,34.96],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]), H298 = (2.9,'kJ/mol','+|-',5.9), S298 = (-144.6,'J/(mol*K)','+|-',8.08), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 546, label = "Cs-CtCdsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 547, label = "Cs-(Cds-O2d)CtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 548, label = "Cs-(Cds-Cd)CtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CtCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 549, label = "Cs-(Cds-Cds)CtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,6.7,8.16,8.92,9.34,9.16,7.14],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (2.99,'kcal/mol','+|-',0.26), S298 = (-34.8,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CtCdCsCs BOZZELLI =3D Cs/Cs3/Cd + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 550, label = "Cs-(Cds-Cdd)CtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 551, label = "Cs-(Cds-Cdd-O2d)CtCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-CsCsCd(CCO)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 552, label = "Cs-(Cds-Cdd-Cd)CtCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 553, label = "Cs-CbCdsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 554, label = "Cs-(Cds-O2d)CbCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 555, label = "Cs-(Cds-Cd)CbCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 556, label = "Cs-(Cds-Cds)CbCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,6.7,8.16,8.92,9.34,9.16,7.14],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (2.99,'kcal/mol','+|-',0.26), S298 = (-34.8,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCdCsCs BOZZELLI =3D Cs/Cs3/Cb + (Cs/Cs3/Cd - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 557, label = "Cs-(Cds-Cdd)CbCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 558, label = "Cs-(Cds-Cdd-O2d)CbCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-CsCsCd(CCO)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 559, label = "Cs-(Cds-Cdd-Cd)CbCsCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 560, label = "Cs-CtCtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.57,5.98,7.51,8.37,9,9.02,8.34],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (1.16,'kcal/mol','+|-',0.26), S298 = (-35.26,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CtCtCsCs BOZZELLI =3D Cs/Cs3/Ct + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 1835, label = "Cs-(CtN3t)(CtN3t)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} {6,T} 3 Ct u0 {1,S} {7,T} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 N3t u0 {2,T} 7 N3t u0 {3,T} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (28.4,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 561, label = "Cs-CbCtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.57,5.98,7.51,8.37,9,9.02,8.34],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (1.16,'kcal/mol','+|-',0.26), S298 = (-35.26,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCtCsCs BOZZELLI =3D Cs/Cs3/Cb + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 562, label = "Cs-CbCbCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.57,5.98,7.51,8.37,9,9.02,8.34],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (1.16,'kcal/mol','+|-',0.26), S298 = (-35.26,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCbCsCs BENSON""", longDesc = u""" """, ) entry( index = 563, label = "Cs-CdsCdsCdsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 564, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cs u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-CsCsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 565, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cs u0 {1,S} 6 C u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([35.99,39.53,39.94,39.09,36.71,34.8,32.51],'J/(mol*K)','+|-',[5.08,5.08,5.08,5.08,5.08,5.08,5.08]), H298 = (19.9,'kJ/mol','+|-',4.33), S298 = (-150.69,'J/(mol*K)','+|-',5.92), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 566, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cs u0 {1,S} 6 Cd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 567, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cs u0 {1,S} 6 Cdd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 568, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 569, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 570, label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 571, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cs u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 572, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 573, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 574, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 575, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 576, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 577, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 578, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 579, label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cs u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 580, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cs u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.32,5.86,7.57,8.54,9.22,9.36,8.45],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (2.54,'kcal/mol','+|-',0.26), S298 = (-33.96,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CdCdCdCs BOZZELLI =3D Cs/Cs2/Cd2 + (Cs/Cs3/Cd - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 581, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cs u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 582, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 583, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 584, label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cs u0 {1,S} 6 Cd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 585, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 586, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 587, label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 Cd u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 588, label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 589, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cs u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 590, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cs u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 591, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cs u0 {1,S} 9 O2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cs u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cs u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cs u0 {1,S} 9 S2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 592, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cs u0 {1,S} 9 C u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 593, label = "Cs-CtCdsCdsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 594, label = "Cs-(Cds-O2d)(Cds-O2d)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 595, label = "Cs-(Cds-O2d)(Cds-Cd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 596, label = "Cs-(Cds-O2d)(Cds-Cds)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 597, label = "Cs-(Cds-O2d)(Cds-Cdd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 598, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 599, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 600, label = "Cs-(Cds-Cd)(Cds-Cd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 601, label = "Cs-(Cds-Cds)(Cds-Cds)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 602, label = "Cs-(Cds-Cdd)(Cds-Cds)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 603, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 604, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 605, label = "Cs-(Cds-Cdd)(Cds-Cdd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 606, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 607, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 608, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Cs u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 609, label = "Cs-CbCdsCdsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 610, label = "Cs-(Cds-O2d)(Cds-O2d)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 611, label = "Cs-(Cds-O2d)(Cds-Cd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 612, label = "Cs-(Cds-O2d)(Cds-Cds)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 613, label = "Cs-(Cds-O2d)(Cds-Cdd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 614, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 615, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 616, label = "Cs-(Cds-Cd)(Cds-Cd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 617, label = "Cs-(Cds-Cds)(Cds-Cds)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 618, label = "Cs-(Cds-Cdd)(Cds-Cds)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 619, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 620, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 621, label = "Cs-(Cds-Cdd)(Cds-Cdd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 622, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 623, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cs u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 624, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cs u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 625, label = "Cs-CtCtCdsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 626, label = "Cs-(Cds-O2d)CtCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 627, label = "Cs-(Cds-Cd)CtCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CtCtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 628, label = "Cs-(Cds-Cds)CtCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,7.36,8.89,9.58,9.76,9.16,7.25],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (5.1,'kcal/mol','+|-',0.26), S298 = (-34.88,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CtCtCdCs BOZZELLI =3D Cs/Cd2/Cs2 + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 629, label = "Cs-(Cds-Cdd)CtCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtCtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 630, label = "Cs-(Cds-Cdd-O2d)CtCtCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtCtCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 631, label = "Cs-(Cds-Cdd-Cd)CtCtCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtCtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 632, label = "Cs-CbCtCdsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 633, label = "Cs-(Cds-O2d)CbCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 634, label = "Cs-(Cds-Cd)CbCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 635, label = "Cs-(Cds-Cds)CbCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,7.36,8.89,9.58,9.76,9.16,7.25],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (5.1,'kcal/mol','+|-',0.26), S298 = (-34.88,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCtCdCs BOZZELLI =3D Cs/Cb/Cd/Cs2 + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 636, label = "Cs-(Cds-Cdd)CbCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 637, label = "Cs-(Cds-Cdd-O2d)CbCtCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCtCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 638, label = "Cs-(Cds-Cdd-Cd)CbCtCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 C u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,7.36,8.89,9.58,9.76,9.16,7.25],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (5.1,'kcal/mol','+|-',0.26), S298 = (-34.88,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCtCdCs BOZZELLI =3D Cs/Cb/Cd/Cs2 + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 639, label = "Cs-CbCbCdsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 640, label = "Cs-(Cds-O2d)CbCbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 641, label = "Cs-(Cds-Cd)CbCbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 642, label = "Cs-(Cds-Cds)CbCbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.99,7.36,8.89,9.58,9.76,9.16,7.25],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (5.1,'kcal/mol','+|-',0.26), S298 = (-34.88,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCbCdCs BOZZELLI =3D Cs/Cs2/Cb2 + (Cs/Cs3/Cd - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 643, label = "Cs-(Cds-Cdd)CbCbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 644, label = "Cs-(Cds-Cdd-O2d)CbCbCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCbCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 645, label = "Cs-(Cds-Cdd-Cd)CbCbCs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCbCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 646, label = "Cs-CtCtCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.37,8.11,9.55,10.1,10.03,9.36,6.65],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (6.23,'kcal/mol','+|-',0.26), S298 = (-35.34,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CtCtCtCs BOZZELLI =3D Cs/Cs2/Ct2 + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 647, label = "Cs-CbCtCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.37,8.11,9.55,10.1,10.03,9.36,6.65],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (6.23,'kcal/mol','+|-',0.26), S298 = (-35.34,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCtCtCs BOZZELLI =3D Cs/Cs2/Cb/Ct + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 648, label = "Cs-CbCbCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.37,8.11,9.55,10.1,10.03,9.36,6.65],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (6.43,'kcal/mol','+|-',0.26), S298 = (-35.34,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCbCtCs BOZZELLI =3D Cs/Cs2/Cb2 + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 649, label = "Cs-CbCbCbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.37,8.11,9.55,10.1,10.03,9.36,6.65],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (6.23,'kcal/mol','+|-',0.26), S298 = (-35.34,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCbCbCs BOZZELLI =3D Cs/Cs2/Cb2 + (Cs/Cs3/Cb - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 650, label = "Cs-CdsCdsCdsCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 651, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 O2d u0 {2,D} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-CsCsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 652, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 CO u0 {1,S} {9,D} 5 Cd u0 {1,S} {6,D} 6 C u0 {5,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} 9 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 653, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 CO u0 {1,S} {9,D} 5 Cd u0 {1,S} {6,D} 6 Cd u0 {5,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} 9 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 654, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 CO u0 {1,S} {9,D} 5 Cd u0 {1,S} {6,D} 6 Cdd u0 {5,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} 9 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 655, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 656, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 657, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)(Cds-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 CO u0 {1,S} {9,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {1,S} {7,D} 6 C u0 {4,D} 7 C u0 {5,D} 8 O2d u0 {2,D} 9 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([42.49,50.96,52.27,50.54,45.33,41.1,35.7],'J/(mol*K)','+|-',[5.08,5.08,5.08,5.08,5.08,5.08,5.08]), H298 = (25.2,'kJ/mol','+|-',4.33), S298 = (-168.67,'J/(mol*K)','+|-',5.92), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 658, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 CO u0 {1,S} {9,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {1,S} {7,D} 6 Cd u0 {4,D} 7 Cd u0 {5,D} 8 O2d u0 {2,D} 9 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 659, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 CO u0 {1,S} {9,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {1,S} {7,D} 6 Cdd u0 {4,D} 7 Cd u0 {5,D} 8 O2d u0 {2,D} 9 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 660, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CO u0 {1,S} {8,D} 4 CO u0 {1,S} {9,D} 5 Cd u0 {1,S} {7,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {5,D} 8 O2d u0 {3,D} 9 O2d u0 {4,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 661, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CO u0 {1,S} {8,D} 4 CO u0 {1,S} {9,D} 5 Cd u0 {1,S} {7,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {5,D} 8 O2d u0 {3,D} 9 O2d u0 {4,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 662, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 CO u0 {1,S} {9,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {1,S} {7,D} 6 Cdd u0 {4,D} 7 Cdd u0 {5,D} 8 O2d u0 {2,D} 9 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 663, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 664, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 665, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 666, label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)(Cds-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 C u0 {3,D} 7 C u0 {4,D} 8 C u0 {5,D} 9 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 667, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 Cd u0 {5,D} 9 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 668, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 Cdd u0 {5,D} 9 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 669, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CO u0 {1,S} {9,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {4,D} 8 Cd u0 {5,D} 9 O2d u0 {3,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 670, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CO u0 {1,S} {9,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {4,D} 8 Cd u0 {5,D} 9 O2d u0 {3,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 671, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cd u0 {3,D} 7 Cdd u0 {4,D} 8 Cdd u0 {5,D} 9 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 672, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CO u0 {1,S} {9,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {5,D} 9 O2d u0 {4,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 673, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CO u0 {1,S} {9,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {5,D} 9 O2d u0 {4,D} 10 O2d u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 674, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CO u0 {1,S} {9,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {5,D} 9 O2d u0 {4,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 675, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 Cdd u0 {5,D} 9 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 676, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} 12 O2d u0 {8,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 677, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} 12 C u0 {8,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 678, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 679, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 CO u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 O2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 680, label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)(Cds-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 C u0 {2,D} 7 C u0 {3,D} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 681, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} """, thermo = u'Cs-CsCsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 682, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 Cdd u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 683, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 684, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 685, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cdd u0 {4,D} 9 Cdd u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 686, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 687, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 688, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {4,D} 9 Cd u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 689, label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} 9 Cdd u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 690, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cd u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} 12 O2d u0 {8,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 691, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cd u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} 12 C u0 {8,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 692, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cd u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cd u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} 12 S2d u0 {8,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cd u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} 12 C u0 {8,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cd u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 693, label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cd u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 694, label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} 9 Cdd u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 695, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} 12 O2d u0 {8,D} 13 O2d u0 {9,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 696, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} 12 O2d u0 {8,D} 13 C u0 {9,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 697, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} 12 C u0 {8,D} 13 C u0 {9,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 698, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 O2d u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} 13 C u0 {9,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} 12 S2d u0 {8,D} 13 S2d u0 {9,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} 12 S2d u0 {8,D} 13 C u0 {9,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} 12 C u0 {8,D} 13 C u0 {9,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 S2d u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} 13 C u0 {9,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 699, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cd u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 Cdd u0 {5,D} {13,D} 10 C u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} 13 C u0 {9,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 700, label = "Cs-CtCdsCdsCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 701, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Ct u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 702, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Ct u0 {1,S} 6 C u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 703, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Ct u0 {1,S} 6 Cd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 704, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Ct u0 {1,S} 6 Cdd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 705, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 706, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 707, label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 708, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Ct u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 709, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Ct u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 710, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 711, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 712, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Ct u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 713, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 714, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 715, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 716, label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Ct u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 717, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Ct u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 718, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Ct u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 719, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 720, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 721, label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Ct u0 {1,S} 6 Cd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 722, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 723, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 724, label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 Cd u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 725, label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Ct u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 726, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Ct u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 727, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Ct u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 728, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Ct u0 {1,S} 9 O2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Ct u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Ct u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Ct u0 {1,S} 9 S2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 729, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Ct u0 {1,S} 9 C u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 730, label = "Cs-CbCdsCdsCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 731, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cb u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 732, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cb u0 {1,S} 6 C u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 733, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cb u0 {1,S} 6 Cd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 734, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cb u0 {1,S} 6 Cdd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 735, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 736, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 737, label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cb u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 738, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cb u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 739, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cb u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 740, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 741, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 742, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cb u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 743, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 744, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 745, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 746, label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cb u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 747, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cb u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 748, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cb u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 749, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 750, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 751, label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cb u0 {1,S} 6 Cd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 752, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 753, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 754, label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 Cd u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 755, label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cb u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 756, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cb u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 757, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cb u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 758, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cb u0 {1,S} 9 O2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cb u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cb u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cb u0 {1,S} 9 S2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 759, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 Cb u0 {1,S} 9 C u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 760, label = "Cs-CtCtCdsCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 761, label = "Cs-(Cds-O2d)(Cds-O2d)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 762, label = "Cs-(Cds-O2d)(Cds-Cd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 763, label = "Cs-(Cds-O2d)(Cds-Cds)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 764, label = "Cs-(Cds-O2d)(Cds-Cdd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 765, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 766, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 767, label = "Cs-(Cds-Cd)(Cds-Cd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 768, label = "Cs-(Cds-Cds)(Cds-Cds)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.61,7.3,8.97,9.69,9.84,9.42,7.36],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (5.48,'kcal/mol','+|-',0.26), S298 = (-34.5,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CtCtCdCd BOZZELLI =3D Cs/Cs/Cd/Ct2 + (Cs/Cs3/Cd - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 769, label = "Cs-(Cds-Cdd)(Cds-Cds)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 770, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 771, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 772, label = "Cs-(Cds-Cdd)(Cds-Cdd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 773, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Ct u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 774, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Ct u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Ct u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Ct u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 775, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 Ct u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 776, label = "Cs-CbCtCdsCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 777, label = "Cs-(Cds-O2d)(Cds-O2d)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 778, label = "Cs-(Cds-O2d)(Cds-Cd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 779, label = "Cs-(Cds-O2d)(Cds-Cds)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 780, label = "Cs-(Cds-O2d)(Cds-Cdd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 781, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 782, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 783, label = "Cs-(Cds-Cd)(Cds-Cd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 784, label = "Cs-(Cds-Cds)(Cds-Cds)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.61,7.3,8.97,9.69,9.84,9.42,7.36],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (5.48,'kcal/mol','+|-',0.26), S298 = (-34.5,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCtCdCd BOZZELLI =3D Cs/Cs/Cb/Cd2 + (Cs/Cs3/Ct - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 785, label = "Cs-(Cds-Cdd)(Cds-Cds)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 786, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 787, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 788, label = "Cs-(Cds-Cdd)(Cds-Cdd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 789, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Ct u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 790, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Ct u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Ct u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Ct u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 791, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Ct u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 792, label = "Cs-CbCbCdsCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 793, label = "Cs-(Cds-O2d)(Cds-O2d)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 794, label = "Cs-(Cds-O2d)(Cds-Cd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 795, label = "Cs-(Cds-O2d)(Cds-Cds)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 796, label = "Cs-(Cds-O2d)(Cds-Cdd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 797, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 798, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 799, label = "Cs-(Cds-Cd)(Cds-Cd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 800, label = "Cs-(Cds-Cds)(Cds-Cds)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.61,7.3,8.97,9.69,9.84,9.42,7.36],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]), H298 = (5.48,'kcal/mol','+|-',0.26), S298 = (-34.5,'cal/(mol*K)','+|-',0.13), ), shortDesc = u"""Cs-CbCbCdCd BOZZELLI =3D Cs/Cs/Cb2/Cd + (Cs/Cs3/Cd - Cs/Cs4)""", longDesc = u""" """, ) entry( index = 801, label = "Cs-(Cds-Cdd)(Cds-Cds)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 802, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 803, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 804, label = "Cs-(Cds-Cdd)(Cds-Cdd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 805, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cb u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 806, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cb u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cb u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cb u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 807, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 Cb u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 808, label = "Cs-CtCtCtCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 809, label = "Cs-(Cds-O2d)CtCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CtCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 810, label = "Cs-(Cds-Cds)CtCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 811, label = "Cs-(Cds-Cdd)CtCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 812, label = "Cs-(Cds-Cdd-O2d)CtCtCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtCtCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 813, label = "Cs-(Cds-Cdd-Cd)CtCtCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 814, label = "Cs-CbCtCtCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 815, label = "Cs-(Cds-O2d)CbCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 816, label = "Cs-(Cds-Cd)CbCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 817, label = "Cs-(Cds-Cds)CbCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 818, label = "Cs-(Cds-Cdd)CbCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 819, label = "Cs-(Cds-Cdd-O2d)CbCtCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCtCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 820, label = "Cs-(Cds-Cdd-Cd)CbCtCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 821, label = "Cs-CbCbCtCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 822, label = "Cs-(Cds-O2d)CbCbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 823, label = "Cs-(Cds-Cd)CbCbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 824, label = "Cs-(Cds-Cds)CbCbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 825, label = "Cs-(Cds-Cdd)CbCbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 826, label = "Cs-(Cds-Cdd-O2d)CbCbCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCbCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 827, label = "Cs-(Cds-Cdd-Cd)CbCbCt", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCbCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 828, label = "Cs-CbCbCbCds", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 [Cd,CO] u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 829, label = "Cs-(Cds-O2d)CbCbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 830, label = "Cs-(Cds-Cd)CbCbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 831, label = "Cs-(Cds-Cds)CbCbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 832, label = "Cs-(Cds-Cdd)CbCbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 833, label = "Cs-(Cds-Cdd-O2d)CbCbCb", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCbCb", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 834, label = "Cs-(Cds-Cdd-Cd)CbCbCb", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCbCb', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 835, label = "Cs-CtCtCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 836, label = "Cs-CbCtCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 837, label = "Cs-CbCbCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 838, label = "Cs-CbCbCbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Ct u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 839, label = "Cs-CbCbCbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cb u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)(Cds-Cd)(Cds-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 C u0 {3,D} 7 C u0 {4,D} 8 C u0 {5,D} 9 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 Cdd u0 {5,D} 9 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CS u0 {1,S} {9,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {4,D} 8 Cd u0 {5,D} 9 S2d u0 {3,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CS u0 {1,S} {9,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {4,D} 8 Cd u0 {5,D} 9 S2d u0 {3,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 Cdd u0 {5,D} 9 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 S2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} 12 C u0 {8,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} 12 S2d u0 {8,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cdd u0 {4,D} {12,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} 12 C u0 {8,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 Cd u0 {5,D} 9 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {9,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cd u0 {1,S} {8,D} 6 Cd u0 {3,D} 7 Cdd u0 {4,D} 8 Cdd u0 {5,D} 9 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CS u0 {1,S} {9,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {5,D} 9 S2d u0 {4,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CS u0 {1,S} {9,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {5,D} 9 S2d u0 {4,D} 10 S2d u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CS u0 {1,S} {9,D} 5 Cd u0 {1,S} {8,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 Cd u0 {5,D} 9 S2d u0 {4,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CtCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Ct u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CtCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)(Cds-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cs u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cs u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCtCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1179, label = "Cs-C=SCsCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.14,6.63,7.51,7.98,8.33,8.38,8.24],'cal/(mol*K)'), H298 = (1.36,'kcal/mol'), S298 = (-33.92,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=SCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Ct u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=SCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cs u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=SC=S", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 S2d u0 {2,D} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=SCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cb u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=S(Cds-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 CS u0 {1,S} {9,D} 5 Cd u0 {1,S} {6,D} 6 C u0 {5,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} 9 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=S(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 CS u0 {1,S} {9,D} 5 Cd u0 {1,S} {6,D} 6 Cdd u0 {5,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} 9 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=S(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=S(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=S(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 CS u0 {1,S} {9,D} 5 Cd u0 {1,S} {6,D} 6 Cd u0 {5,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} 9 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)(Cds-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Ct u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Ct u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Ct u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Ct u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CsCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CbCt", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Ct u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CbCb", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cb u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Ct u0 {1,S} 6 C u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cds)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Ct u0 {1,S} 6 Cd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Ct u0 {1,S} 6 Cdd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-Cd)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-S2d)Ct", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Ct u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cs u0 {1,S} 6 C u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cds)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cs u0 {1,S} 6 Cd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cs u0 {1,S} 6 Cdd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-S2d)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-Cd)Cs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cs u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cd)(Cds-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 CS u0 {1,S} {9,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {1,S} {7,D} 6 C u0 {4,D} 7 C u0 {5,D} 8 S2d u0 {2,D} 9 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 CS u0 {1,S} {9,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {1,S} {7,D} 6 Cdd u0 {4,D} 7 Cd u0 {5,D} 8 S2d u0 {2,D} 9 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CS u0 {1,S} {8,D} 4 CS u0 {1,S} {9,D} 5 Cd u0 {1,S} {7,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {5,D} 8 S2d u0 {3,D} 9 S2d u0 {4,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-Cd)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 CS u0 {1,S} {8,D} 4 CS u0 {1,S} {9,D} 5 Cd u0 {1,S} {7,D} 6 Cdd u0 {2,D} {10,D} 7 Cd u0 {5,D} 8 S2d u0 {3,D} 9 S2d u0 {4,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd)(Cds-Cdd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 CS u0 {1,S} {9,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {1,S} {7,D} 6 Cdd u0 {4,D} 7 Cdd u0 {5,D} 8 S2d u0 {2,D} 9 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 CS u0 {1,S} {9,D} 6 Cdd u0 {2,D} {10,D} 7 Cdd u0 {3,D} {11,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cds)(Cds-Cds)", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 CS u0 {1,S} {9,D} 4 Cd u0 {1,S} {6,D} 5 Cd u0 {1,S} {7,D} 6 Cd u0 {4,D} 7 Cd u0 {5,D} 8 S2d u0 {2,D} 9 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cb u0 {1,S} 6 C u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cb u0 {1,S} 6 Cdd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-S2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 Cb u0 {1,S} 6 Cd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCtCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CbCs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 Cs u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)(Cds-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cb u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cb u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cb u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cb u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cb u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Cb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cb u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCbCb", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCbCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 840, label = "Cs-CCCOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 C u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-CsCsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 841, label = "Cs-CsCsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([23.99,31.2,34.89,36.47,36.78,36.05,34.4],'J/(mol*K)','+|-',[3.81,3.81,3.81,3.81,3.81,3.81,3.81]), H298 = (-20.3,'kJ/mol','+|-',3.24), S298 = (-144.38,'J/(mol*K)','+|-',4.44), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 842, label = "Cs-CdsCsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 843, label = "Cs-(Cds-O2d)CsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([28.15,35.17,38.11,38.72,37.49,35.88,33.45],'J/(mol*K)','+|-',[5.16,5.16,5.16,5.16,5.16,5.16,5.16]), H298 = (-10.9,'kJ/mol','+|-',4.39), S298 = (-148.7,'J/(mol*K)','+|-',6.02), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 844, label = "Cs-(Cds-Cd)CsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([29.24,37.61,40.84,41.46,40.06,38.2,35.08],'J/(mol*K)','+|-',[3.81,3.81,3.81,3.81,3.81,3.81,3.81]), H298 = (-14.6,'kJ/mol','+|-',3.24), S298 = (-153.23,'J/(mol*K)','+|-',4.44), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 845, label = "Cs-(Cds-Cds)CsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.63,6.79,7.95,8.4,8.8,8.44,8.44],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]), H298 = (-6.6,'kcal/mol','+|-',0.4), S298 = (-32.56,'cal/(mol*K)','+|-',0.2), ), shortDesc = u"""Cs-OCdCsCs BOZZELLI C/C3/O - (C/C3/H - C/Cb/C2/H), Hf-1 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 846, label = "Cs-(Cds-Cdd)CsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 847, label = "Cs-(Cds-Cdd-O2d)CsCsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.39,9.66,10.03,10.07,9.64,9.26,8.74],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]), H298 = (-9.725,'kcal/mol','+|-',0.4), S298 = (-36.5,'cal/(mol*K)','+|-',0.2), ), shortDesc = u"""{C/CCO/O/C2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 848, label = "Cs-(Cds-Cdd-Cd)CsCsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 849, label = "Cs-OsCtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 O2s u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 850, label = "Cs-CbCsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.63,6.79,7.95,8.4,8.8,8.44,8.44],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]), H298 = (-6.6,'kcal/mol','+|-',0.4), S298 = (-32.56,'cal/(mol*K)','+|-',0.2), ), shortDesc = u"""Cs-OCbCsCs BOZZELLI C/C3/O - (C/C3/H - C/Cb/C2/H), Hf-1 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 851, label = "Cs-CdsCdsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 852, label = "Cs-(Cds-O2d)(Cds-O2d)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-CsCsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 853, label = "Cs-(Cds-O2d)(Cds-Cd)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([33.75,42.15,45.09,44.95,41.74,38.55,34.46],'J/(mol*K)','+|-',[4.3,4.3,4.3,4.3,4.3,4.3,4.3]), H298 = (-3.9,'kJ/mol','+|-',3.66), S298 = (-158.3,'J/(mol*K)','+|-',5.02), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 854, label = "Cs-(Cds-O2d)(Cds-Cds)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 855, label = "Cs-(Cds-O2d)(Cds-Cdd)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 856, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 857, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 858, label = "Cs-(Cds-Cd)(Cds-Cd)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 859, label = "Cs-(Cds-Cds)(Cds-Cds)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.61,5.98,7.51,8.37,9,9.02,8.34],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]), H298 = (-8.01,'kcal/mol','+|-',0.4), S298 = (-34.34,'cal/(mol*K)','+|-',0.2), ), shortDesc = u"""Cs-OCdCdCs Hf jwb 697 S,Cp from C/Cd2/C2""", longDesc = u""" """, ) entry( index = 860, label = "Cs-(Cds-Cdd)(Cds-Cds)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 861, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 862, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 863, label = "Cs-(Cds-Cdd)(Cds-Cdd)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 864, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 865, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 866, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 O2s u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 867, label = "Cs-CtCdsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 868, label = "Cs-(Cds-O2d)CtCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 869, label = "Cs-(Cds-Cd)CtCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CtCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 870, label = "Cs-(Cds-Cds)CtCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 871, label = "Cs-(Cds-Cdd)CtCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 872, label = "Cs-(Cds-Cdd-O2d)CtCsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 873, label = "Cs-(Cds-Cdd-Cd)CtCsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 874, label = "Cs-CbCdsCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 875, label = "Cs-(Cds-O2d)CbCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 876, label = "Cs-(Cds-Cd)CbCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 877, label = "Cs-(Cds-Cds)CbCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 878, label = "Cs-(Cds-Cdd)CbCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 879, label = "Cs-(Cds-Cdd-O2d)CbCsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 880, label = "Cs-(Cds-Cdd-Cd)CbCsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 881, label = "Cs-CtCtCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 882, label = "Cs-CbCtCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 883, label = "Cs-CbCbCsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 884, label = "Cs-CdsCdsCdsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 885, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 O2s u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-CsCsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 886, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 O2s u0 {1,S} 6 C u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 887, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 O2s u0 {1,S} 6 Cd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 888, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 O2s u0 {1,S} 6 Cdd u0 {4,D} 7 O2d u0 {2,D} 8 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 889, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 890, label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {7,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 891, label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 O2s u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([36.85,46.04,49,48.85,45.61,42.23,37.25],'J/(mol*K)','+|-',[4.09,4.09,4.09,4.09,4.09,4.09,4.09]), H298 = (3,'kJ/mol','+|-',3.49), S298 = (-160.69,'J/(mol*K)','+|-',4.77), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 892, label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 O2s u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 893, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 O2s u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 894, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 O2s u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 895, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CO u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 O2s u0 {1,S} 7 Cd u0 {4,D} 8 O2d u0 {3,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 896, label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 O2s u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 897, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 898, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 899, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CO u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 900, label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 O2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 901, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 O2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} """, thermo = u'Cs-CsCsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 902, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 O2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 903, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsCsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 904, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 905, label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 O2s u0 {1,S} 6 Cd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 906, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 O2s u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 O2d u0 {6,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 907, label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 O2s u0 {1,S} 8 Cd u0 {4,D} 9 O2d u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 908, label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 O2s u0 {1,S} 8 Cd u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 909, label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 910, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 O2s u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 O2d u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 911, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 O2s u0 {1,S} 9 O2d u0 {5,D} 10 O2d u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 912, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 O2s u0 {1,S} 9 O2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 913, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 O2s u0 {1,S} 9 C u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 914, label = "Cs-CtCdsCdsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 915, label = "Cs-(Cds-O2d)(Cds-O2d)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 916, label = "Cs-(Cds-O2d)(Cds-Cd)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 917, label = "Cs-(Cds-O2d)(Cds-Cds)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 918, label = "Cs-(Cds-O2d)(Cds-Cdd)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 919, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 920, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 921, label = "Cs-(Cds-Cd)(Cds-Cd)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 922, label = "Cs-(Cds-Cds)(Cds-Cds)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 923, label = "Cs-(Cds-Cdd)(Cds-Cds)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 924, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 925, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 926, label = "Cs-(Cds-Cdd)(Cds-Cdd)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 927, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 928, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 929, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 O2s u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 930, label = "Cs-CbCdsCdsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 931, label = "Cs-(Cds-O2d)(Cds-O2d)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 932, label = "Cs-(Cds-O2d)(Cds-Cd)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 933, label = "Cs-(Cds-O2d)(Cds-Cds)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 934, label = "Cs-(Cds-O2d)(Cds-Cdd)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 935, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 936, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 937, label = "Cs-(Cds-Cd)(Cds-Cd)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 938, label = "Cs-(Cds-Cds)(Cds-Cds)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 939, label = "Cs-(Cds-Cdd)(Cds-Cds)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 940, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 941, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 942, label = "Cs-(Cds-Cdd)(Cds-Cdd)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 943, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 944, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 945, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 O2s u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 946, label = "Cs-CtCtCdsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 947, label = "Cs-(Cds-O2d)CtCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 948, label = "Cs-(Cds-Cd)CtCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CtCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 949, label = "Cs-(Cds-Cds)CtCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 950, label = "Cs-(Cds-Cdd)CtCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 951, label = "Cs-(Cds-Cdd-O2d)CtCtOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 952, label = "Cs-(Cds-Cdd-Cd)CtCtOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 953, label = "Cs-CbCtCdsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 954, label = "Cs-(Cds-O2d)CbCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 955, label = "Cs-(Cds-Cd)CbCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 956, label = "Cs-(Cds-Cds)CbCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 957, label = "Cs-(Cds-Cdd)CbCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 958, label = "Cs-(Cds-Cdd-O2d)CbCtOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 959, label = "Cs-(Cds-Cdd-Cd)CbCtOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 960, label = "Cs-CbCbCdsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 [Cd,CO] u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbCbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 961, label = "Cs-(Cds-O2d)CbCbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 962, label = "Cs-(Cds-Cd)CbCbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbCbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 963, label = "Cs-(Cds-Cds)CbCbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 964, label = "Cs-(Cds-Cdd)CbCbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbCbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 965, label = "Cs-(Cds-Cdd-O2d)CbCbOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 966, label = "Cs-(Cds-Cdd-Cd)CbCbOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbCbOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 967, label = "Cs-CtCtCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 968, label = "Cs-CbCtCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtCtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 969, label = "Cs-CbCbCtOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 970, label = "Cs-CbCbCbOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 971, label = "Cs-CCOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-CsCsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 972, label = "Cs-CsCsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([27.88,43.75,51.85,54,50.77,45.94,38.31],'J/(mol*K)','+|-',[5.77,5.77,5.77,5.77,5.77,5.77,5.77]), H298 = (-69.2,'kJ/mol','+|-',4.92), S298 = (-163.77,'J/(mol*K)','+|-',6.74), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 973, label = "Cs-CdsCsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 974, label = "Cs-(Cds-O2d)CsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-CsCsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 975, label = "Cs-(Cds-Cd)CsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([27.95,42.92,51.33,54.81,53.92,49.73,41.11],'J/(mol*K)','+|-',[5.77,5.77,5.77,5.77,5.77,5.77,5.77]), H298 = (-62.8,'kJ/mol','+|-',4.92), S298 = (-170.44,'J/(mol*K)','+|-',6.74), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 976, label = "Cs-(Cds-Cds)CsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-CsCsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 977, label = "Cs-(Cds-Cdd)CsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 978, label = "Cs-(Cds-Cdd-O2d)CsOsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 979, label = "Cs-(Cds-Cdd-Cd)CsOsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 980, label = "Cs-CdsCdsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 981, label = "Cs-(Cds-O2d)(Cds-O2d)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-CsCsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 982, label = "Cs-(Cds-O2d)(Cds-Cd)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 983, label = "Cs-(Cds-O2d)(Cds-Cds)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 984, label = "Cs-(Cds-O2d)(Cds-Cdd)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 985, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 986, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 987, label = "Cs-(Cds-Cd)(Cds-Cd)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([30.08,45.85,54.7,58.39,57.78,53.65,44.31],'J/(mol*K)','+|-',[5.77,5.77,5.77,5.77,5.77,5.77,5.77]), H298 = (-55.7,'kJ/mol','+|-',4.92), S298 = (-179.76,'J/(mol*K)','+|-',6.74), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 988, label = "Cs-(Cds-Cds)(Cds-Cds)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-CsCsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 989, label = "Cs-(Cds-Cdd)(Cds-Cds)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 990, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 991, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 992, label = "Cs-(Cds-Cdd)(Cds-Cdd)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 993, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 O2s u0 {1,S} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 994, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 O2s u0 {1,S} 7 O2s u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 995, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsOs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 O2s u0 {1,S} 7 O2s u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 996, label = "Cs-CtCsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 997, label = "Cs-CtCdsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 998, label = "Cs-(Cds-O2d)CtOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 999, label = "Cs-(Cds-Cd)CtOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CtOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1000, label = "Cs-(Cds-Cds)CtOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1001, label = "Cs-(Cds-Cdd)CtOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1002, label = "Cs-(Cds-Cdd-O2d)CtOsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1003, label = "Cs-(Cds-Cdd-Cd)CtOsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1004, label = "Cs-CtCtOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1005, label = "Cs-CbCsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1006, label = "Cs-CbCdsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1007, label = "Cs-(Cds-O2d)CbOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1008, label = "Cs-(Cds-Cd)CbOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1009, label = "Cs-(Cds-Cds)CbOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1010, label = "Cs-(Cds-Cdd)CbOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1011, label = "Cs-(Cds-Cdd-O2d)CbOsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1012, label = "Cs-(Cds-Cdd-Cd)CbOsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1013, label = "Cs-CbCtOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1014, label = "Cs-CbCbOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1015, label = "Cs-COsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-CsOsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1016, label = "Cs-CsOsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.33,6.19,7.25,7.7,8.2,8.24,8.24],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]), H298 = (-19,'kcal/mol','+|-',0.4), S298 = (-33.56,'cal/(mol*K)','+|-',0.2), ), shortDesc = u"""Cs-OOOCs BOZZELLI est !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1017, label = "Cs-CdsOsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)OsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1018, label = "Cs-(Cds-O2d)OsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-CsOsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1019, label = "Cs-(Cds-Cd)OsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)OsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1020, label = "Cs-(Cds-Cds)OsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-CsOsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1021, label = "Cs-(Cds-Cdd)OsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)OsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1022, label = "Cs-(Cds-Cdd-O2d)OsOsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)OsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1023, label = "Cs-(Cds-Cdd-Cd)OsOsOs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 O2s u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)OsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1024, label = "Cs-CtOsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)OsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1025, label = "Cs-CbOsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = u'Cs-(Cds-Cds)OsOsOs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1026, label = "Cs-OsOsOsOs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 O2s u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.33,6.13,7.25,7.7,8.2,8.24,8.24],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]), H298 = (-23,'kcal/mol','+|-',0.4), S298 = (-35.56,'cal/(mol*K)','+|-',0.2), ), shortDesc = u"""Cs-OOOO BOZZELLI est !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1027, label = "Cs-COsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CsOsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1028, label = "Cs-CsOsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.25,7.1,8.81,9.55,10.31,11.05,11.05],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]), H298 = (-16,'kcal/mol','+|-',0.24), S298 = (-12.07,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cs-OOCsH BENSON Hf, BOZZELLI C/C3/H - C/C2/O/H !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1029, label = "Cs-CdsOsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)OsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1030, label = "Cs-(Cds-O2d)OsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-CsOsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1031, label = "Cs-(Cds-Cd)OsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)OsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1032, label = "Cs-(Cds-Cds)OsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-CsOsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1033, label = "Cs-(Cds-Cdd)OsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)OsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1034, label = "Cs-(Cds-Cdd-O2d)OsOsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cds)OsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1035, label = "Cs-(Cds-Cdd-Cd)OsOsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 O2s u0 {1,S} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)OsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1036, label = "Cs-CtOsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)OsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1037, label = "Cs-CbOsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)OsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-COsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 O2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1452, label = "Cs-CsOsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 O2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.37,10.32,11.1,11.3,11.3,11.21,11.6],'cal/(mol*K)'), H298 = (-11.1,'kcal/mol'), S298 = (-16.14,'cal/(mol*K)'), ), shortDesc = u"""CAC CBS-QB3 1DHR calc""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsOsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 O2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtOsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 O2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbOsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 O2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CCOsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 O2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1453, label = "Cs-CsCsOsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.16,10.15,10.69,10.52,9.74,9.01,8.34],'cal/(mol*K)'), H298 = (-11.26,'kcal/mol'), S298 = (-39.73,'cal/(mol*K)'), ), shortDesc = u"""CAC CBS-QB3 1DHR calc""", longDesc = u""" """, ) entry( index = 1467, label = "Cs-COsOsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = u'Cs-CsOsOsSs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1465, label = "Cs-CsOsOsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 O2s u0 {1,S} 4 O2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.65,8.43,9.23,9.47,9.43,9.2,8.89],'cal/(mol*K)'), H298 = (-21.41,'kcal/mol'), S298 = (-36.7,'cal/(mol*K)'), ), shortDesc = u"""CAC calc 1D-HR""", longDesc = u""" """, ) entry( index = 1038, label = "Cs-CCOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CsCsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1039, label = "Cs-CsCsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([21.99,29.03,34.22,37.78,41.96,44.27,47.11],'J/(mol*K)','+|-',[3.32,3.32,3.32,3.32,3.32,3.32,3.32]), H298 = (-25.1,'kJ/mol','+|-',2.83), S298 = (-52.05,'J/(mol*K)','+|-',3.88), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1040, label = "Cs-CdsCsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1041, label = "Cs-(Cds-O2d)CsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.47,6.82,8.45,9.17,10.24,10.8,11.02],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]), H298 = (-6,'kcal/mol','+|-',0.24), S298 = (-11.1,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cs-OCOCsH BOZZELLI""", longDesc = u""" """, ) entry( index = 1042, label = "Cs-(Cds-Cd)CsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([29.84,38.86,43.83,46.37,48.34,49.06,49.94],'J/(mol*K)','+|-',[3.74,3.74,3.74,3.74,3.74,3.74,3.74]), H298 = (-24,'kJ/mol','+|-',3.19), S298 = (-61.06,'J/(mol*K)','+|-',4.36), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1043, label = "Cs-(Cds-Cds)CsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.47,6.82,8.45,9.17,10.24,10.8,11.02],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]), H298 = (-6,'kcal/mol','+|-',0.24), S298 = (-11.1,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cs-OCdCsH BOZZELLI""", longDesc = u""" """, ) entry( index = 1044, label = "Cs-(Cds-Cdd)CsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1045, label = "Cs-(Cds-Cdd-O2d)CsOsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.2,8.49,9.33,9.92,10.5,10.92,11.71],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]), H298 = (-8.37,'kcal/mol','+|-',0.24), S298 = (-13.04,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""{C/CCO/O/C/H} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 1046, label = "Cs-(Cds-Cdd-Cd)CsOsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1047, label = "Cs-CdsCdsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1048, label = "Cs-(Cds-O2d)(Cds-O2d)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 CO u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} 7 O2d u0 {3,D} """, thermo = u'Cs-CsCsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1049, label = "Cs-(Cds-O2d)(Cds-Cd)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1050, label = "Cs-(Cds-O2d)(Cds-Cds)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)CsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1051, label = "Cs-(Cds-O2d)(Cds-Cdd)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1052, label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1053, label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CO u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1054, label = "Cs-(Cds-Cd)(Cds-Cd)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([29.82,38.47,43.27,45.7,47.5,48.09,48.78],'J/(mol*K)','+|-',[3.64,3.64,3.64,3.64,3.64,3.64,3.64]), H298 = (-17.4,'kJ/mol','+|-',3.1), S298 = (-64.14,'J/(mol*K)','+|-',4.24), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1055, label = "Cs-(Cds-Cds)(Cds-Cds)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.21,6.6,8.26,9.05,10.23,10.86,11.04],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]), H298 = (-6.67,'kcal/mol','+|-',0.24), S298 = (-10.42,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cs-OCdCdH BOZZELLI""", longDesc = u""" """, ) entry( index = 1056, label = "Cs-(Cds-Cdd)(Cds-Cds)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1057, label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 O2d u0 {4,D} """, thermo = u'Cs-(Cds-Cdd-O2d)CsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1058, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1059, label = "Cs-(Cds-Cdd)(Cds-Cdd)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1060, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 O2s u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 O2d u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1061, label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 O2s u0 {1,S} 7 H u0 {1,S} 8 O2d u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1062, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 O2s u0 {1,S} 7 H u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1063, label = "Cs-CtCsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CsOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1064, label = "Cs-CtCdsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1065, label = "Cs-(Cds-O2d)CtOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1066, label = "Cs-(Cds-Cd)CtOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CtOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1067, label = "Cs-(Cds-Cds)CtOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1068, label = "Cs-(Cds-Cdd)CtOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CtOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1069, label = "Cs-(Cds-Cdd-O2d)CtOsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1070, label = "Cs-(Cds-Cdd-Cd)CtOsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CtOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1071, label = "Cs-CtCtOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1072, label = "Cs-CbCsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.47,6.82,8.45,9.17,10.24,10.8,11.02],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]), H298 = (-6,'kcal/mol','+|-',0.24), S298 = (-11.1,'cal/(mol*K)','+|-',0.12), ), shortDesc = u"""Cs-OCbCsH BOZZELLI =3D C/Cd/C/H/O Jul 91""", longDesc = u""" """, ) entry( index = 1073, label = "Cs-CbCdsOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 [Cd,CO] u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CbOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1074, label = "Cs-(Cds-O2d)CbOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1075, label = "Cs-(Cds-Cd)CbOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = u'Cs-(Cds-Cds)CbOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1076, label = "Cs-(Cds-Cds)CbOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1077, label = "Cs-(Cds-Cdd)CbOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)CbOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1078, label = "Cs-(Cds-Cdd-O2d)CbOsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1079, label = "Cs-(Cds-Cdd-Cd)CbOsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 O2s u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)CbOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1080, label = "Cs-CbCtOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)CtOsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1081, label = "Cs-CbCbOsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 O2s u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1082, label = "Cs-COsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-CsOsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1083, label = "Cs-CsOsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([25.01,31.9,37.45,41.88,48.53,53.31,60.53],'J/(mol*K)','+|-',[1.43,1.43,1.43,1.43,1.43,1.43,1.43]), H298 = (-34.3,'kJ/mol','+|-',1.22), S298 = (37.65,'J/(mol*K)','+|-',1.67), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1084, label = "Cs-CdsOsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 [Cd,CO] u0 {1,S} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)OsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1085, label = "Cs-(Cds-O2d)OsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CO u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([26.75,34.37,40.77,45.37,51.2,54.96,60.79],'J/(mol*K)','+|-',[4.34,4.34,4.34,4.34,4.34,4.34,4.34]), H298 = (-19.8,'kJ/mol','+|-',3.7), S298 = (31.54,'J/(mol*K)','+|-',5.06), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1086, label = "Cs-(Cds-Cd)OsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([28.42,35.65,40.62,44.31,49.79,53.92,60.6],'J/(mol*K)','+|-',[3.38,3.38,3.38,3.38,3.38,3.38,3.38]), H298 = (-26.6,'kJ/mol','+|-',2.88), S298 = (34.59,'J/(mol*K)','+|-',3.95), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1087, label = "Cs-(Cds-Cds)OsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.12,6.86,8.32,9.49,11.22,12.48,14.4],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-6.76,'kcal/mol','+|-',0.2), S298 = (9.8,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cs-OCdHH BOZZELLI Hf PEDLEY c*ccoh C/C/Cd/H2""", longDesc = u""" """, ) entry( index = 1088, label = "Cs-(Cds-Cdd)OsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = u'Cs-(Cds-Cdd-Cd)OsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1089, label = "Cs-(Cds-Cdd-O2d)OsHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.15,8.67,9.75,10.65,11.93,12.97,14.86],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-8.68,'kcal/mol','+|-',0.2), S298 = (8.43,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""{C/CCO/O/H2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""", longDesc = u""" """, ) entry( index = 1090, label = "Cs-(Cds-Cdd-Cd)OsHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 O2s u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = u'Cs-(Cds-Cds)OsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1091, label = "Cs-CtOsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.12,6.86,8.32,9.49,11.22,12.48,14.4],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-6.76,'kcal/mol','+|-',0.2), S298 = (9.8,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""Cs-OCtHH BOZZELLI assigned C/Cd/H2/O""", longDesc = u""" """, ) entry( index = 1092, label = "Cs-CbOsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 O2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = u'Cs-(Cds-Cds)OsHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CCCSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 C u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1175, label = "Cs-CsCsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.41,7.05,8.02,8.53,8.87,8.85,8.57],'cal/(mol*K)'), H298 = (-0.49,'kcal/mol'), S298 = (-34.44,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsCsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CsCsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CsCsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-SsCtCsCs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 S2s u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsCdsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 Cd u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)(Cds-Cd)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cds)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cdd)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cs u0 {1,S} 7 S2s u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCdsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cd u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CtCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CtCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CtCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtCsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CtCsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCdsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cd u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CbCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CbCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CbCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CbCsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCtCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCtCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCbCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsCdsCdsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 Cd u0 {1,S} 4 Cd u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 S2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 S2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cd u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 S2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 Cd u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 S2s u0 {1,S} 6 Cd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 S2s u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 S2s u0 {1,S} 8 Cd u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 S2s u0 {1,S} 8 Cd u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cd u0 {1,S} {8,D} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} 8 Cdd u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 S2s u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 S2d u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 S2s u0 {1,S} 9 S2d u0 {5,D} 10 S2d u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 S2s u0 {1,S} 9 S2d u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {8,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 Cdd u0 {4,D} {11,D} 8 S2s u0 {1,S} 9 C u0 {5,D} 10 C u0 {6,D} 11 C u0 {7,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCdsCdsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cd u0 {1,S} 4 Cd u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)(Cds-Cd)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cds)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cdd)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Ct u0 {1,S} 7 S2s u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCdsCdsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cd u0 {1,S} 4 Cd u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)(Cds-Cd)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cds)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cdd)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 Cb u0 {1,S} 7 S2s u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCtCdsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Cd u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CtCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CtCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CtCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtCtSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CtCtSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 Ct u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCtCdsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Cd u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CbCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CbCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CbCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCtSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CbCtSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Ct u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCbCdsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cd u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CbCbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CbCbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CbCbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbCbSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CbCbSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 Cb u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCtCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCtCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCbCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCbCbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCsCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)(Cds-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 S2s u0 {1,S} 6 C u0 {3,D} 7 C u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 S2s u0 {1,S} 6 Cdd u0 {3,D} 7 Cdd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 C u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {7,S} 2 Cd u0 {1,S} {5,D} 3 Cd u0 {1,S} {6,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 Cdd u0 {3,D} {10,D} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} 10 S2d u0 {6,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)(Cds-Cds)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 S2s u0 {1,S} 6 Cdd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 S2s u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {7,D} 5 Cdd u0 {2,D} {9,D} 6 S2s u0 {1,S} 7 Cd u0 {4,D} 8 S2d u0 {3,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)(Cds-Cds)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {8,D} 3 Cd u0 {1,S} {6,D} 4 Cd u0 {1,S} {7,D} 5 S2s u0 {1,S} 6 Cd u0 {3,D} 7 Cd u0 {4,D} 8 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CtSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Ct u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SC=SSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 S2s u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 S2s u0 {1,S} 6 C u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cds)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 S2s u0 {1,S} 6 Cd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 CS u0 {1,S} {8,D} 4 Cd u0 {1,S} {6,D} 5 S2s u0 {1,S} 6 Cdd u0 {4,D} 7 S2d u0 {2,D} 8 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-S2d)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=S(Cds-Cdd-Cd)S2s", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {6,S} 2 Cd u0 {1,S} {5,D} 3 CS u0 {1,S} {7,D} 4 CS u0 {1,S} {8,D} 5 Cdd u0 {2,D} {9,D} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cb u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CbSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SCtSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)CsSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 Cs u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CCSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1202, label = "Cs-CsCsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.17,8.72,9.4,9.63,9.55,9.29,8.67],'cal/(mol*K)'), H298 = (-1.34,'kcal/mol'), S298 = (-36.66,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsCsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CsSsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CsSsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsCdsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 Cd u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)(Cds-Cd)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cds)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cdd)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 S2s u0 {1,S} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 S2s u0 {1,S} 7 S2s u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 S2s u0 {1,S} 7 S2s u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCdsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cd u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CtSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CtSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CtSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtSsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CtSsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCtSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCdsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cd u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CbSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CbSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CbSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbSsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CbSsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCtSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCbSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)SsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CSsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1203, label = "Cs-CsSsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.45,9.59,10.49,10.71,10.42,9.94,8.92],'cal/(mol*K)'), H298 = (-1.8,'kcal/mol'), S298 = (-38.19,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsSsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)SsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)SsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)SsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)SsSsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)SsSsSs", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 S2s u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtSsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbSsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SSsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-SsSsSsSs", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 S2s u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 S2s u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CSsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1173, label = "Cs-CsSsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.82,9.05,9.73,10.14,10.63,10.97,11.44],'cal/(mol*K)'), H298 = (-3.3,'kcal/mol'), S298 = (-14.59,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsSsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)SsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)SsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)SsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)SsSsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)SsSsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 S2s u0 {1,S} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtSsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbSsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SSsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CCSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1169, label = "Cs-CsCsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.01,6.7,7.96,8.83,9.89,10.51,11.27],'cal/(mol*K)'), H298 = (-1.98,'kcal/mol'), S298 = (-11.89,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1170, label = "Cs-CdsCsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.41,9.85,10.93,11.28,11.37,11.41,11.61],'cal/(mol*K)'), H298 = (-2.15,'kcal/mol'), S298 = (-15.26,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CsSsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CsSsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cs u0 {1,S} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CdsCdsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 Cd u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)(Cds-Cd)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)(Cds-Cds)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cds)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 Cd u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)(Cds-Cdd)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cd u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} 7 Cdd u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 S2s u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 S2d u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 S2s u0 {1,S} 7 H u0 {1,S} 8 S2d u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {6,S} {7,S} 2 Cd u0 {1,S} {4,D} 3 Cd u0 {1,S} {5,D} 4 Cdd u0 {2,D} {8,D} 5 Cdd u0 {3,D} {9,D} 6 S2s u0 {1,S} 7 H u0 {1,S} 8 C u0 {4,D} 9 C u0 {5,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1171, label = "Cs-CtCsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.08,6.78,7.93,8.72,9.73,10.35,11.19],'cal/(mol*K)'), H298 = (0.72,'kcal/mol'), S298 = (-11.64,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCdsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Cd u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CtSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CtSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CtSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CtSsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CtSsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Ct u0 {1,S} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CtCtSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1172, label = "Cs-CbCsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.38,7.96,9.03,9.69,10.45,10.89,11.47],'cal/(mol*K)'), H298 = (-1.66,'kcal/mol'), S298 = (-13.65,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCdsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cd u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)CbSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)CbSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)CbSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)CbSsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)CbSsH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 Cb u0 {1,S} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCtSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CbCbSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCbSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cb u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SC=SSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 CS u0 {1,S} {7,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1174, label = "Cs-C=SCsSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Cs u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([7.53,10.09,11.25,11.65,11.76,11.74,11.77],'cal/(mol*K)'), H298 = (-3.49,'kcal/mol'), S298 = (-15.86,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=SCtSsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 Ct u0 {1,S} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cd)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 C u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-Cd)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 C u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cdd-S2d)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {5,S} {6,S} 2 Cd u0 {1,S} {4,D} 3 CS u0 {1,S} {7,D} 4 Cdd u0 {2,D} {8,D} 5 S2s u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} 8 S2d u0 {4,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-C=S(Cds-Cds)SsH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {7,D} 3 Cd u0 {1,S} {6,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {3,D} 7 S2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-CSsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 C u0 {1,S} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1163, label = "Cs-CsSsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cs u0 {1,S} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.11,6.59,7.97,9.15,10.99,12.33,14.32],'cal/(mol*K)'), H298 = (-4.94,'kcal/mol'), S298 = (9.92,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1164, label = "Cs-CdsSsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.17,9.71,10.55,11.14,12.11,12.95,14.43],'cal/(mol*K)'), H298 = (-5.07,'kcal/mol'), S298 = (6.75,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cd)SsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cds)SsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd)SsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cd u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 Cdd u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-S2d)SsHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 S2d u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "Cs-(Cds-Cdd-Cd)SsHH", group = """ 1 * Cs u0 {2,S} {4,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} 3 Cdd u0 {2,D} {7,D} 4 S2s u0 {1,S} 5 H u0 {1,S} 6 H u0 {1,S} 7 C u0 {3,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1165, label = "Cs-CtSsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Ct u0 {1,S} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.19,7.02,8.42,9.51,11.14,12.34,14.18],'cal/(mol*K)'), H298 = (-2.69,'kcal/mol'), S298 = (9.75,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1166, label = "Cs-CbSsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 Cb u0 {1,S} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.3,7.96,9.19,10.12,11.53,12.59,14.26],'cal/(mol*K)'), H298 = (-5.04,'kcal/mol'), S298 = (8.26,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1168, label = "Cs-C=SSsHH", group = """ 1 * Cs u0 {2,S} {3,S} {4,S} {5,S} 2 CS u0 {1,S} {6,D} 3 S2s u0 {1,S} 4 H u0 {1,S} 5 H u0 {1,S} 6 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.23,10.43,11.42,11.97,12.76,13.43,14.63],'cal/(mol*K)'), H298 = (-6.13,'kcal/mol'), S298 = (5.73,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1093, label = "O", group = """ 1 * O u0 """, thermo = u'O2s-CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1094, label = "O2d", group = """ 1 * O2d u0 """, thermo = u'O2d-Cd', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1095, label = "O2d-Cd", group = """ 1 * O2d u0 {2,D} 2 CO u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""In this case the C is treated as the central atom""", longDesc = u""" """, ) entry( index = 1096, label = "O2d-O2d", group = """ 1 * O2d u0 {2,D} 2 O2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.5,3.575,3.685,3.8,3.99,4.12,4.29],'cal/(mol*K)'), H298 = (14.01,'kcal/mol'), S298 = (24.085,'cal/(mol*K)'), ), shortDesc = u"""A. Vandeputte""", longDesc = u""" """, ) entry( index = 1943, label = "O2d-N3d", group = """ 1 * O2d u0 {2,D} 2 N3d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1944, label = "O2d-N5dc", group = """ 1 * O2d u0 {2,D} 2 N5dc u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1097, label = "O2s", group = """ 1 * O2s u0 """, thermo = u'O2s-(Cds-Cd)(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1945, label = "O2s-N", group = """ 1 * O2s u0 {2,S} 2 N u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1935, label = "O2s-CN", group = """ 1 * O2s u0 {2,S} {3,S} 2 C u0 {1,S} 3 N u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1874, label = "O2s-CsN3s", group = """ 1 * O2s u0 {2,S} {3,S} 2 N3s u0 {1,S} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.5,3.6,4,4.3,4.7,4.8,4.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (-9.2,'kcal/mol','+|-',1.3), S298 = (7.2,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1875, label = "O2s-CsN3d", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 N3d u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1847, label = "O2s-Cs(N3dOd)", group = """ 1 * O2s u0 {2,S} {4,S} 2 N3d u0 {1,S} {3,D} 3 O2d u0 {2,D} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.6,11.3,11.9,12.6,13.6,14.3,14.8],'cal/(mol*K)','+|-',[0.8,0.8,0.8,0.8,0.8,0.8,0.8]), H298 = (-4.8,'kcal/mol','+|-',1.1), S298 = (40,'cal/(mol*K)','+|-',1), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1877, label = "O2s-CdN3d", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cd u0 {1,S} 3 N3d u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1878, label = "O2s-(Cd-Cd)(N3dOd)", group = """ 1 Cd u0 {2,S} {4,D} {5,S} 2 * O2s u0 {1,S} {3,S} 3 N3d u0 {2,S} {6,D} 4 Cd u0 {1,D} 5 R u0 {1,S} 6 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.1,11.7,12.2,12.7,13.5,14.1,14.9],'cal/(mol*K)','+|-',[0.7,0.7,0.7,0.7,0.7,0.7,0.7]), H298 = (-5.3,'kcal/mol','+|-',0.9), S298 = (39.5,'cal/(mol*K)','+|-',0.9), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1876, label = "O2s-CsN5d", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 N5dc u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1848, label = "O2s-Cs(N5dOdOs)", group = """ 1 N5dc u0 {2,S} {3,D} {4,S} 2 * O2s u0 {1,S} {5,S} 3 O2d u0 {1,D} 4 O2s u0 {1,S} 5 Cs u0 {2,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.2,13.9,15.4,16.6,18.4,19.3,19.9],'cal/(mol*K)','+|-',[0.8,0.8,0.8,0.8,0.8,0.8,0.8]), H298 = (-19.1,'kcal/mol','+|-',1.1), S298 = (45.3,'cal/(mol*K)','+|-',1), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1879, label = "O2s-CdN5d", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cd u0 {1,S} 3 N5dc u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1880, label = "O2s-(Cd-CdHH)(N5dOdOs)", group = """ 1 N5dc u0 {3,S} {4,D} {5,S} 2 Cd u0 {3,S} {6,D} {7,S} 3 * O2s u0 {1,S} {2,S} 4 O2d u0 {1,D} 5 O2s u0 {1,S} 6 Cd u0 {2,D} 7 R u0 {2,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([12.4,14.2,15.7,16.9,18.5,19.3,20.1],'cal/(mol*K)','+|-',[0.8,0.8,0.8,0.8,0.8,0.8,0.8]), H298 = (-18.4,'kcal/mol','+|-',1.1), S298 = (45.4,'cal/(mol*K)','+|-',1), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1936, label = "O2s-ON", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 N u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1881, label = "O2s-OsN3s", group = """ 1 * O2s u0 {2,S} {3,S} 2 N3s u0 {1,S} 3 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.3,4.9,5.6,6.3,7,7.1,6.5],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (5.3,'kcal/mol','+|-',1.3), S298 = (6.9,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1882, label = "O2s-OsN3d", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 N3d u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1883, label = "O2s-O2s(N3dOd)", group = """ 1 * O2s u0 {2,S} {4,S} 2 N3d u0 {1,S} {3,D} 3 O2d u0 {2,D} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.7,12.9,13.6,14.2,15,15.5,16],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (15.2,'kcal/mol','+|-',1.3), S298 = (40.7,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1937, label = "O2s-NN", group = """ 1 * O2s u0 {2,S} {3,S} 2 N u0 {1,S} 3 N u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1884, label = "O2s-N3sN3s", group = """ 1 * O2s u0 {2,S} {3,S} 2 N3s u0 {1,S} 3 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.8,4.6,5.1,5.2,5.2,4.9,4.3],'cal/(mol*K)','+|-',[1.6,1.6,1.6,1.6,1.6,1.6,1.6]), H298 = (5.7,'kcal/mol','+|-',2.2), S298 = (6.8,'cal/(mol*K)','+|-',2.1), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1885, label = "O2s-N3sN3d", group = """ 1 * O2s u0 {2,S} {3,S} 2 N3s u0 {1,S} 3 N3d u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1886, label = "O2s-N3s(N3dOd)", group = """ 1 * O2s u0 {2,S} {4,S} 2 N3d u0 {1,S} {3,D} 3 O2d u0 {2,D} 4 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.2,11.5,12.4,13,13.9,14.3,14.8],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (10.8,'kcal/mol','+|-',1.3), S298 = (40.8,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1098, label = "O2s-HH", group = """ 1 * O2s u0 {2,S} {3,S} 2 H u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.03,8.19,8.42,8.68,9.26,9.86,11.26],'cal/(mol*K)'), H298 = (-57.8,'kcal/mol','+|-',0.01), S298 = (46.51,'cal/(mol*K)','+|-',0.002), ), shortDesc = u"""O-HH WATER. !!!Using NIST value for H2O, S(group) = S(H2O) + Rln(2)""", longDesc = u""" """, ) entry( index = 1099, label = "O2s-OsH", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.21,5.72,6.17,6.66,7.15,7.61,8.43],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]), H298 = (-16.3,'kcal/mol','+|-',0.14), S298 = (27.83,'cal/(mol*K)','+|-',0.07), ), shortDesc = u"""O-OH SANDIA 1/2*H2O2""", longDesc = u""" """, ) entry( index = 1100, label = "O2s-OsOs", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2.2,3.64,4.2,4.34,4.62,4.9,4.9],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (8.85,'kcal/mol','+|-',0.16), S298 = (9.4,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""O-OO LAY 1997=20 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1101, label = "O2s-CH", group = """ 1 * O2s u0 {2,S} {3,S} 2 C u0 {1,S} 3 H u0 {1,S} """, thermo = u'O2s-CsH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1102, label = "O2s-CtH", group = """ 1 * O2s u0 {2,S} {3,S} 2 Ct u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.3,4.5,4.82,5.23,6.02,6.61,7.44],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-37.9,'kcal/mol','+|-',0.16), S298 = (29.1,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""O-CtH BENSON (Assigned O-CsH)""", longDesc = u""" """, ) entry( index = 1103, label = "O2s-CdsH", group = """ 1 * O2s u0 {2,S} {3,S} 2 [Cd,CO] u0 {1,S} 3 H u0 {1,S} """, thermo = u'O2s-(Cds-Cd)H', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1104, label = "O2s-(Cds-O2d)H", group = """ 1 * O2s u0 {2,S} {3,S} 2 CO u0 {1,S} {4,D} 3 H u0 {1,S} 4 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([19.07,19.8,20.85,22.07,24.57,26.95,31.66],'J/(mol*K)','+|-',[2.54,2.54,2.54,2.54,2.54,2.54,2.54]), H298 = (-165.2,'kJ/mol','+|-',2.16), S298 = (125.32,'J/(mol*K)','+|-',2.96), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1105, label = "O2s-(Cds-Cd)H", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cd u0 {1,S} {4,D} 3 H u0 {1,S} 4 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([24.6,30.3,32.52,33.15,33.29,33.55,34.97],'J/(mol*K)','+|-',[4.18,4.18,4.18,4.18,4.18,4.18,4.18]), H298 = (-188.1,'kJ/mol','+|-',3.56), S298 = (106.3,'J/(mol*K)','+|-',4.87), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1106, label = "O2s-CsH", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([19.07,19.8,20.85,22.07,24.57,26.95,31.66],'J/(mol*K)','+|-',[2.54,2.54,2.54,2.54,2.54,2.54,2.54]), H298 = (-165.2,'kJ/mol','+|-',2.16), S298 = (125.32,'J/(mol*K)','+|-',2.96), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1107, label = "O2s-CbH", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cb u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.3,4.5,4.82,5.23,6.02,6.61,7.44],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-37.9,'kcal/mol','+|-',0.16), S298 = (29.1,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""O-CbH BENSON (Assigned O-CsH)""", longDesc = u""" """, ) entry( index = 1460, label = "O2s-CSH", group = """ 1 * O2s u0 {2,S} {3,S} 2 CS u0 {1,S} {4,D} 3 H u0 {1,S} 4 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.98,8.35,9.48,10.38,11.59,12.26,12.99],'cal/(mol*K)'), H298 = (-31.38,'kcal/mol'), S298 = (32.08,'cal/(mol*K)'), ), shortDesc = u"""CAC calc 1D-HR""", longDesc = u""" """, ) entry( index = 1108, label = "O2s-OsC", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 C u0 {1,S} """, thermo = u'O2s-OsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1109, label = "O2s-OsCt", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.9,4.31,4.6,4.84,5.32,5.8,5.8],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (7,'kcal/mol','+|-',0.3), S298 = (10.8,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""O-OCb Hf JWB plot S,Cp assigned O/O/Cd !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1110, label = "O2s-OsCds", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 [Cd,CO] u0 {1,S} """, thermo = u'O2s-O2s(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1111, label = "O2s-O2s(Cds-O2d)", group = """ 1 * O2s u0 {2,S} {3,S} 2 CO u0 {1,S} {4,D} 3 O2s u0 {1,S} 4 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.53,5.02,5.79,6.08,6.54,6.49,6.49],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (-23.22,'kcal/mol','+|-',0.3), S298 = (9.11,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""O-OCO jwl cbsQ 99 cqcho=20 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1112, label = "O2s-O2s(Cds-Cd)", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cd u0 {1,S} {4,D} 3 O2s u0 {1,S} 4 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.5,3.87,3.95,4.15,4.73,4.89,4.89],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (1.64,'kcal/mol','+|-',0.3), S298 = (10.12,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""O-OCd WESTMORELAND S,Cp LAY'9405 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1113, label = "O2s-OsCs", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.9,4.31,4.6,4.84,5.32,5.8,5.8],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]), H298 = (-5.4,'kcal/mol','+|-',0.3), S298 = (8.54,'cal/(mol*K)','+|-',0.15), ), shortDesc = u"""O-OCs LAY 1997 !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1114, label = "O2s-OsCb", group = """ 1 * O2s u0 {2,S} {3,S} 2 O2s u0 {1,S} 3 Cb u0 {1,S} """, thermo = u'O2s-O2s(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1115, label = "O2s-CC", group = """ 1 * O2s u0 {2,S} {3,S} 2 C u0 {1,S} 3 C u0 {1,S} """, thermo = u'O2s-(Cds-Cd)(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1116, label = "O2s-CtCt", group = """ 1 * O2s u0 {2,S} {3,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} """, thermo = u'O2s-(Cds-Cd)(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1117, label = "O2s-CtCds", group = """ 1 * O2s u0 {2,S} {3,S} 2 Ct u0 {1,S} 3 [Cd,CO] u0 {1,S} """, thermo = u'O2s-(Cds-Cd)(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1118, label = "O2s-Ct(Cds-O2d)", group = """ 1 * O2s u0 {2,S} {3,S} 2 CO u0 {1,S} {4,D} 3 Ct u0 {1,S} 4 O2d u0 {2,D} """, thermo = u'O2s-(Cds-Cd)(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1119, label = "O2s-Ct(Cds-Cd)", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cd u0 {1,S} {4,D} 3 Ct u0 {1,S} 4 C u0 {2,D} """, thermo = u'O2s-(Cds-Cd)(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1120, label = "O2s-CtCs", group = """ 1 * O2s u0 {2,S} {3,S} 2 Ct u0 {1,S} 3 Cs u0 {1,S} """, thermo = u'O2s-Cs(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1873, label = "O2s-Cs(CtN3t)", group = """ 1 * O2s u0 {2,S} {4,S} 2 Ct u0 {1,S} {3,T} 3 N3t u0 {2,T} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([9.1,9.8,10.6,11.2,12.3,13,13.8],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (10,'kcal/mol','+|-',1.3), S298 = (39.1,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1121, label = "O2s-CtCb", group = """ 1 * O2s u0 {2,S} {3,S} 2 Ct u0 {1,S} 3 Cb u0 {1,S} """, thermo = u'O2s-(Cds-Cd)(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1122, label = "O2s-CdsCds", group = """ 1 * O2s u0 {2,S} {3,S} 2 [Cd,CO] u0 {1,S} 3 [Cd,CO] u0 {1,S} """, thermo = u'O2s-(Cds-Cd)(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1123, label = "O2s-(Cds-O2d)(Cds-O2d)", group = """ 1 * O2s u0 {2,S} {3,S} 2 CO u0 {1,S} {4,D} 3 CO u0 {1,S} {5,D} 4 O2d u0 {2,D} 5 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([18.4,11.55,6.97,3.72,-0.53,-2.57,-1.41],'J/(mol*K)','+|-',[6.51,6.51,6.51,6.51,6.51,6.51,6.51]), H298 = (-46.4,'kJ/mol','+|-',5.54), S298 = (80.8,'J/(mol*K)','+|-',7.59), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1124, label = "O2s-(Cds-O2d)(Cds-Cd)", group = """ 1 * O2s u0 {2,S} {3,S} 2 CO u0 {1,S} {5,D} 3 Cd u0 {1,S} {4,D} 4 C u0 {3,D} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([20.02,19.61,18.5,17.71,17.02,16.49,15.33],'J/(mol*K)','+|-',[8.17,8.17,8.17,8.17,8.17,8.17,8.17]), H298 = (-100.6,'kJ/mol','+|-',6.96), S298 = (38.43,'J/(mol*K)','+|-',9.53), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1125, label = "O2s-(Cds-Cd)(Cds-Cd)", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cd u0 {1,S} 3 Cd u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.4,3.7,3.7,3.8,4.4,4.6,4.8],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-19.61,'kcal/mol','+|-',0.19), S298 = (10,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""O-CdCd BOZZELLI""", longDesc = u""" """, ) entry( index = 1126, label = "O2s-CdsCs", group = """ 1 * O2s u0 {2,S} {3,S} 2 [Cd,CO] u0 {1,S} 3 Cs u0 {1,S} """, thermo = u'O2s-Cs(Cds-Cd)', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1127, label = "O2s-Cs(Cds-O2d)", group = """ 1 * O2s u0 {2,S} {3,S} 2 CO u0 {1,S} {4,D} 3 Cs u0 {1,S} 4 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.49,9.94,9.96,10.7,12.71,14.71,18],'J/(mol*K)','+|-',[3.15,3.15,3.15,3.15,3.15,3.15,3.15]), H298 = (-102.2,'kJ/mol','+|-',2.69), S298 = (45.71,'J/(mol*K)','+|-',3.68), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1128, label = "O2s-Cs(Cds-Cd)", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cd u0 {1,S} {4,D} 3 Cs u0 {1,S} 4 C u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([19.07,23.32,25.26,25.92,25.5,24.52,22.72],'J/(mol*K)','+|-',[3.47,3.47,3.47,3.47,3.47,3.47,3.47]), H298 = (-123.9,'kJ/mol','+|-',2.96), S298 = (18.91,'J/(mol*K)','+|-',4.05), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = -1, label = "O2s-CdsCb", group = """ 1 * O2s u0 {2,S} {3,S} 2 [Cd,CO] u0 {1,S} 3 Cb u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "O2s-Cb(Cds-O2d)", group = """ 1 * O2s u0 {2,S} {3,S} 2 CO u0 {1,S} {4,D} 3 Cb u0 {1,S} 4 O2d u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "O2s-Cb(Cds-Cd)", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cd u0 {1,S} {4,D} 3 Cb u0 {1,S} 4 C u0 {2,D} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1129, label = "O2s-CsCs", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([14.7,13.4,13.58,14.54,16.71,18.29,20.17],'J/(mol*K)','+|-',[2.44,2.44,2.44,2.44,2.44,2.44,2.44]), H298 = (-98.6,'kJ/mol','+|-',2.08), S298 = (38.61,'J/(mol*K)','+|-',2.85), ), shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""", longDesc = u""" Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452, DOI: 10.1002/chem.201301381 """, ) entry( index = 1130, label = "O2s-CsCb", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.4,3.7,3.7,3.8,4.4,4.6,4.6],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-22.6,'kcal/mol','+|-',0.19), S298 = (9.7,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""O-CbCs REID, PRAUSNITZ and SHERWOOD !!!WARNING! Cp1500 value taken as Cp1000""", longDesc = u""" """, ) entry( index = 1131, label = "O2s-CbCb", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([1.19,-0.24,-0.72,-0.51,0.43,1.36,1.75],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]), H298 = (-18.77,'kcal/mol','+|-',0.19), S298 = (13.59,'cal/(mol*K)','+|-',0.1), ), shortDesc = u"""O-CbCb CHERN 1/97 Hf PEDLEY, Mopac""", longDesc = u""" """, ) entry( index = 1461, label = "O2s-Cs(Cds-S2d)", group = """ 1 * O2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 CS u0 {1,S} {4,D} 4 S2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.56,6.31,7,7.61,8.52,8.99,9.29],'cal/(mol*K)'), H298 = (-14.54,'kcal/mol'), S298 = (10.02,'cal/(mol*K)'), ), shortDesc = u"""CAC CBS-QB3 1D-HR""", longDesc = u""" """, ) entry( index = 1600, label = "Si", group = """ 1 * Si u0 """, thermo = u'Cs-HHHH', shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1400, label = "S", group = """ 1 * S u0 """, thermo = u'S2s-CsCs', shortDesc = u"""""", longDesc = u""" """, ) entry( index = -1, label = "S2d", group = """ 1 * S2d u0 """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1160, label = "S2d-Cd", group = """ 1 * S2d u0 {2,D} 2 CS u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1161, label = "S2d-S2d", group = """ 1 * S2d u0 {2,D} 2 S2d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.9,4.08,4.2,4.27,4.35,4.39,4.43],'cal/(mol*K)'), H298 = (22.82,'kcal/mol'), S298 = (26.89,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2009""", longDesc = u""" """, ) entry( index = -1, label = "S2s", group = """ 1 * S2s u0 """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1132, label = "S2s-HH", group = """ 1 * S2s u0 {2,S} {3,S} 2 H u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.15,8.48,8.85,9.26,10.08,10.82,12.1],'cal/(mol*K)'), H298 = (-5.37,'kcal/mol'), S298 = (50.52,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "S2s-CH", group = """ 1 * S2s u0 {2,S} {3,S} 2 C u0 {1,S} 3 H u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1133, label = "S2s-CsH", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.17,6.22,6.4,6.65,7.18,7.65,8.45],'cal/(mol*K)'), H298 = (5.05,'kcal/mol'), S298 = (33.68,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1134, label = "S2s-CdH", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cd u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.01,6.82,7.28,7.55,7.9,8.18,8.7],'cal/(mol*K)'), H298 = (4.19,'kcal/mol'), S298 = (32.23,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1135, label = "S2s-CtH", group = """ 1 * S2s u0 {2,S} {3,S} 2 Ct u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.22,6.87,7.26,7.55,7.94,8.24,8.73],'cal/(mol*K)'), H298 = (6.27,'kcal/mol'), S298 = (31.59,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1136, label = "S2s-CbH", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cb u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.65,7.07,7.33,7.5,7.79,8.05,8.53],'cal/(mol*K)'), H298 = (3.91,'kcal/mol'), S298 = (31.98,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1462, label = "S2s-COH", group = """ 1 * S2s u0 {2,S} {3,S} 2 CO u0 {1,S} {4,D} 3 H u0 {1,S} 4 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.05,9.1,9.95,10.65,11.62,12.26,13.25],'cal/(mol*K)'), H298 = (-21.06,'kcal/mol'), S298 = (35.41,'cal/(mol*K)'), ), shortDesc = u"""CAC calc 1D-HR""", longDesc = u""" """, ) entry( index = 1153, label = "S2s-C=SH", group = """ 1 * S2s u0 {2,S} {3,S} 2 CS u0 {1,S} {4,D} 3 H u0 {1,S} 4 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.01,6.82,7.28,7.55,7.9,8.18,8.7],'cal/(mol*K)'), H298 = (4.19,'kcal/mol'), S298 = (32.23,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1147, label = "S2s-SsH", group = """ 1 * S2s u0 {2,S} {3,S} 2 S2s u0 {1,S} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.75,6.48,7.02,7.43,8.03,8.43,9],'cal/(mol*K)'), H298 = (1.97,'kcal/mol'), S298 = (31.73,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1152, label = "S2s-SsSs", group = """ 1 * S2s u0 {2,S} {3,S} 2 S2s u0 {1,S} 3 S2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.19,6.32,6.38,6.44,6.47,6.39,5.95],'cal/(mol*K)'), H298 = (3.03,'kcal/mol'), S298 = (11.18,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "S2s-SsC", group = """ 1 * S2s u0 {2,S} {3,S} 2 S2s u0 {1,S} 3 C u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1148, label = "S2s-SsCs", group = """ 1 * S2s u0 {2,S} {3,S} 2 S2s u0 {1,S} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.59,5.73,5.79,5.8,5.74,5.65,5.43],'cal/(mol*K)'), H298 = (6.99,'kcal/mol'), S298 = (12.61,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1149, label = "S2s-SsCd", group = """ 1 * S2s u0 {2,S} {3,S} 2 S2s u0 {1,S} 3 Cd u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.41,5.92,6.11,6.14,5.98,5.74,5.25],'cal/(mol*K)'), H298 = (7.62,'kcal/mol'), S298 = (12.13,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1150, label = "S2s-SsCt", group = """ 1 * S2s u0 {2,S} {3,S} 2 S2s u0 {1,S} 3 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.3,5.8,5.94,5.94,5.77,5.57,5.24],'cal/(mol*K)'), H298 = (11.93,'kcal/mol'), S298 = (12.73,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1151, label = "S2s-SsCb", group = """ 1 * S2s u0 {2,S} {3,S} 2 S2s u0 {1,S} 3 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.71,5.93,5.98,5.92,5.67,5.38,4.78],'cal/(mol*K)'), H298 = (7.09,'kcal/mol'), S298 = (11.38,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1159, label = "S2s-C=SSs", group = """ 1 * S2s u0 {2,S} {3,S} 2 S2s u0 {1,S} 3 CS u0 {1,S} {4,D} 4 S2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.25,5.49,5.65,5.74,5.74,5.65,5.37],'cal/(mol*K)'), H298 = (7.9,'kcal/mol'), S298 = (13.34,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = -1, label = "S2s-CC", group = """ 1 * S2s u0 {2,S} {3,S} 2 C u0 {1,S} 3 C u0 {1,S} """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1137, label = "S2s-CsCs", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.8,5.76,5.63,5.51,5.3,5.18,5.07],'cal/(mol*K)'), H298 = (11.41,'kcal/mol'), S298 = (13.72,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1138, label = "S2s-CsCd", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 Cd u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.68,6.98,6.89,6.65,6.16,5.79,5.33],'cal/(mol*K)'), H298 = (9.83,'kcal/mol'), S298 = (11.01,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1463, label = "S2s-CsCO", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 CO u0 {1,S} {4,D} 4 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.78,6.73,7.6,8.34,9.31,9.77,10.14],'cal/(mol*K)'), H298 = (-15.33,'kcal/mol'), S298 = (11.11,'cal/(mol*K)'), ), shortDesc = u"""CAC CBS-QB3 1dhr calc""", longDesc = u""" """, ) entry( index = 1139, label = "S2s-CsCt", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.45,5.71,5.77,5.73,5.57,5.42,5.2],'cal/(mol*K)'), H298 = (12.03,'kcal/mol'), S298 = (13.23,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1140, label = "S2s-CsCb", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.62,5.74,5.7,5.6,5.38,5.22,5],'cal/(mol*K)'), H298 = (10.51,'kcal/mol'), S298 = (12.6,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1141, label = "S2s-CdCd", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cd u0 {1,S} 3 Cd u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.82,6.48,6.62,6.51,6.09,5.74,5.25],'cal/(mol*K)'), H298 = (10.56,'kcal/mol'), S298 = (12.24,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1142, label = "S2s-CdCt", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cd u0 {1,S} 3 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.34,5.99,6.17,6.13,5.88,5.63,5.27],'cal/(mol*K)'), H298 = (12.84,'kcal/mol'), S298 = (12.07,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1143, label = "S2s-CdCb", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cd u0 {1,S} 3 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.91,6.42,6.51,6.38,6,5.65,5.07],'cal/(mol*K)'), H298 = (10.23,'kcal/mol'), S298 = (11.93,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1144, label = "S2s-CtCt", group = """ 1 * S2s u0 {2,S} {3,S} 2 Ct u0 {1,S} 3 Ct u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.61,5.28,5.46,5.47,5.37,5.27,5.11],'cal/(mol*K)'), H298 = (19.93,'kcal/mol'), S298 = (13.38,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1145, label = "S2s-CtCb", group = """ 1 * S2s u0 {2,S} {3,S} 2 Ct u0 {1,S} 3 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.6,5.94,5.94,5.82,5.56,5.35,5.06],'cal/(mol*K)'), H298 = (13.27,'kcal/mol'), S298 = (11.87,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1146, label = "S2s-CbCb", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.27,5.7,5.8,5.74,5.53,5.35,5.09],'cal/(mol*K)'), H298 = (10.52,'kcal/mol'), S298 = (12.32,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1154, label = "S2s-C=SCs", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cs u0 {1,S} 3 CS u0 {1,S} {4,D} 4 S2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.43,5.15,5.65,5.92,6.04,5.97,5.72],'cal/(mol*K)'), H298 = (6.87,'kcal/mol'), S298 = (11.81,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1156, label = "S2s-C=SCt", group = """ 1 * S2s u0 {2,S} {3,S} 2 CS u0 {1,S} {4,D} 3 Ct u0 {1,S} 4 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.39,4.92,5.17,5.28,5.33,5.29,5.19],'cal/(mol*K)'), H298 = (15.16,'kcal/mol'), S298 = (14.06,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1158, label = "S2s-C=SC=S", group = """ 1 * S2s u0 {2,S} {3,S} 2 CS u0 {1,S} {4,D} 3 CS u0 {1,S} {5,D} 4 S2d u0 {2,D} 5 S2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.98,5.28,5.67,6.04,6.51,6.52,5.77],'cal/(mol*K)'), H298 = (12.91,'kcal/mol'), S298 = (12.96,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1155, label = "S2s-C=SCd", group = """ 1 * S2s u0 {2,S} {3,S} 2 Cd u0 {1,S} 3 CS u0 {1,S} {4,D} 4 S2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.13,6.41,7.01,7.14,6.87,6.48,5.84],'cal/(mol*K)'), H298 = (7.78,'kcal/mol'), S298 = (10.23,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1157, label = "S2s-C=SCb", group = """ 1 * S2s u0 {2,S} {3,S} 2 CS u0 {1,S} {4,D} 3 Cb u0 {1,S} 4 S2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.08,5.5,5.68,5.7,5.59,5.42,5.05],'cal/(mol*K)'), H298 = (10.76,'kcal/mol'), S298 = (13.05,'cal/(mol*K)'), ), shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""", longDesc = u""" """, ) entry( index = 1887, label = "N", group = """ 1 * [N1dc,N3s,N3d,N3t,N5sc,N5dc,N5ddc,N5tc] u0 """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1922, label = "N1dc", group = """ 1 * N1dc u0 p2 """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1888, label = "N3s", group = """ 1 * N3s u0 """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1938, label = "N3s-CHH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 C u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1808, label = "N3s-CsHH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.72,6.51,7.32,8.07,9.41,10.47,12.28],'cal/(mol*K)'), H298 = (4.8,'kcal/mol'), S298 = (29.71,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1817, label = "N3s-CbHH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cb u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.72,6.51,7.32,8.07,9.41,10.47,12.28],'cal/(mol*K)'), H298 = (4.8,'kcal/mol'), S298 = (29.71,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1825, label = "N3s-(CO)HH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 CO u0 {1,S} {5,D} 3 H u0 {1,S} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.07,5.74,7.13,8.29,9.96,11.22,14.37],'cal/(mol*K)'), H298 = (-14.9,'kcal/mol'), S298 = (-24.69,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1889, label = "N3s-CdHH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cd u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.7,6.5,7.3,8.1,9.4,10.5,12.3],'cal/(mol*K)'), H298 = (4.8,'kcal/mol'), S298 = (29.7,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1938, label = "N3s-CCH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1809, label = "N3s-CsCsH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.2,5.21,6.13,6.83,7.9,8.65,9.55],'cal/(mol*K)'), H298 = (15.4,'kcal/mol'), S298 = (8.94,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1818, label = "N3s-CbCsH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (14.9,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1820, label = "N3s-CbCbH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cb u0 {1,S} 3 Cb u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (16.3,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1826, label = "N3s-(CO)CsH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 CO u0 {1,S} {5,D} 3 Cs u0 {1,S} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (-4.4,'kcal/mol'), S298 = (3.9,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1828, label = "N3s-(CO)CbH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 CO u0 {1,S} {5,D} 3 Cb u0 {1,S} 4 H u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0.4,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1829, label = "N3s-(CO)(CO)H", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 CO u0 {1,S} {5,D} 3 CO u0 {1,S} {6,D} 4 H u0 {1,S} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (-18.5,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1894, label = "N3s-(CtN3t)CsH", group = """ 1 * N3s u0 {2,S} {4,S} {5,S} 2 Ct u0 {1,S} {3,T} 3 N3t u0 {2,T} 4 Cs u0 {1,S} 5 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.3,11.6,12.8,13.9,15.5,16.7,18.3],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (44.1,'kcal/mol','+|-',1.3), S298 = (40.7,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1899, label = "N3s-(CdCd)CsH", group = """ 1 * N3s u0 {2,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} {4,S} 3 Cd u0 {2,D} 4 R u0 {2,S} 5 Cs u0 {1,S} 6 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.8,6.1,6.4,6.7,7.5,8.1,9.1],'cal/(mol*K)','+|-',[1.3,1.3,1.3,1.3,1.3,1.3,1.3]), H298 = (15.3,'kcal/mol','+|-',1.9), S298 = (8.7,'cal/(mol*K)','+|-',1.7), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1939, label = "N3s-CCC", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 C u0 {1,S} 3 C u0 {1,S} 4 C u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1810, label = "N3s-CsCsCs", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.48,4.56,5.43,5.97,6.56,6.67,6.5],'cal/(mol*K)'), H298 = (24.4,'kcal/mol'), S298 = (-13.46,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1819, label = "N3s-CbCsCs", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cb u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (26.2,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1827, label = "N3s-(CO)CsCs", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 CO u0 {1,S} {5,D} 3 Cs u0 {1,S} 4 Cs u0 {1,S} 5 O2d u0 {2,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1830, label = "N3s-(CO)(CO)Cs", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 CO u0 {1,S} {5,D} 3 CO u0 {1,S} {6,D} 4 Cs u0 {1,S} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (-5.9,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1831, label = "N3s-(CO)(CO)Cb", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 CO u0 {1,S} {5,D} 3 CO u0 {1,S} {6,D} 4 Cb u0 {1,S} 5 O2d u0 {2,D} 6 O2d u0 {3,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (-0.5,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1895, label = "N3s-(CtN3t)CsCs", group = """ 1 * N3s u0 {2,S} {4,S} {5,S} 2 Ct u0 {1,S} {3,T} 3 N3t u0 {2,T} 4 Cs u0 {1,S} 5 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([8.6,9.6,10.5,11.4,12.9,13.8,14.8],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (53.3,'kcal/mol','+|-',1.3), S298 = (21,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1898, label = "N3s-(CdCd)CsCs", group = """ 1 * N3s u0 {2,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} {4,S} 3 Cd u0 {2,D} 4 R u0 {2,S} 5 Cs u0 {1,S} 6 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2.8,2.9,3.3,3.7,4.6,5,5.5],'cal/(mol*K)','+|-',[1.3,1.3,1.3,1.3,1.3,1.3,1.3]), H298 = (25.9,'kcal/mol','+|-',1.9), S298 = (-11,'cal/(mol*K)','+|-',1.7), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1811, label = "N3s-N3sHH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 H u0 {1,S} 3 H u0 {1,S} 4 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.1,7.38,8.43,9.27,10.54,11.52,13.19],'cal/(mol*K)'), H298 = (11.4,'kcal/mol'), S298 = (29.13,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1940, label = "N3s-NCH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 N u0 {1,S} 3 C u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1812, label = "N3s-N3sCsH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 H u0 {1,S} 4 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.82,5.8,6.5,7,7.8,8.3,9],'cal/(mol*K)'), H298 = (20.9,'kcal/mol'), S298 = (9.61,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1814, label = "N3s-N3sCbH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 N3s u0 {1,S} 3 Cb u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (22.1,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1897, label = "N3s-CsH(N3dOd)", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 H u0 {1,S} 4 N3d u0 {1,S} {5,D} 5 O2d u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([10.4,11.9,13.4,14.7,16.6,17.9,19.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (25.2,'kcal/mol','+|-',1.3), S298 = (41.7,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1902, label = "N3s-CsH(N5dOdOs)", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 H u0 {1,S} 4 N5dc u0 {1,S} {5,D} {6,S} 5 O2d u0 {4,D} 6 O2s u0 {4,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([13.1,15.5,17.6,19.2,21.4,22.8,24.4],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (8.4,'kcal/mol','+|-',1.3), S298 = (45.3,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1901, label = "N3s-(CdCd)HN3s", group = """ 1 * N3s u0 {2,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} {4,S} 3 Cd u0 {2,D} 4 R u0 {2,S} 5 H u0 {1,S} 6 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.5,5.4,6.5,7.3,8.5,9.1,9.9],'cal/(mol*K)','+|-',[1.1,1.1,1.1,1.1,1.1,1.1,1.1]), H298 = (20.5,'kcal/mol','+|-',1.5), S298 = (6.6,'cal/(mol*K)','+|-',1.4), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1940, label = "N3s-NCC", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 N u0 {1,S} 3 C u0 {1,S} 4 C u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1893, label = "N3s-NCsCs", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 N u0 {1,S} 3 Cs u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (29.2,'kcal/mol'), S298 = (-13.8,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1813, label = "N3s-CsCsN3s", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.7,4.9,5.8,6.3,6.8,6.8,6.7],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (26.8,'kcal/mol','+|-',1.3), S298 = (-14.5,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1896, label = "N3s-CsCs(N3dOd)", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 N3d u0 {1,S} {5,D} 5 O2d u0 {4,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([9.4,10.5,11.5,12.4,13.8,14.6,15.3],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (32.6,'kcal/mol','+|-',1.3), S298 = (19.3,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1903, label = "N3s-CsCs(N5dOdOs)", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 N5dc u0 {1,S} {5,D} {6,S} 5 O2d u0 {4,D} 6 O2s u0 {4,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([11.5,13.4,15.2,16.7,18.8,20,21.1],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (16.7,'kcal/mol','+|-',1.3), S298 = (25.8,'cal/(mol*K)','+|-',1.2), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1941, label = "N3s-NCdCs", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 N u0 {1,S} 3 Cd u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1900, label = "N3s-(CdCd)CsN3s", group = """ 1 * N3s u0 {2,S} {5,S} {6,S} 2 Cd u0 {1,S} {3,D} {4,S} 3 Cd u0 {2,D} 4 R u0 {2,S} 5 Cs u0 {1,S} 6 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.2,4.2,4.4,4.8,5.4,5.7,6],'cal/(mol*K)','+|-',[1.1,1.1,1.1,1.1,1.1,1.1,1.1]), H298 = (30.3,'kcal/mol','+|-',1.5), S298 = (-13.2,'cal/(mol*K)','+|-',1.4), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1891, label = "N3s-CsHOs", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 H u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([5.2,6.2,7,7.7,8.7,9.4,10.5],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]), H298 = (20.4,'kcal/mol','+|-',1.4), S298 = (8.1,'cal/(mol*K)','+|-',1.3), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1892, label = "N3s-CsCsOs", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 Cs u0 {1,S} 3 Cs u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.3,5.1,5.7,6.2,7,7.3,7.5],'cal/(mol*K)','+|-',[0.8,0.8,0.8,0.8,0.8,0.8,0.8]), H298 = (26.6,'kcal/mol','+|-',1.2), S298 = (-12.7,'cal/(mol*K)','+|-',1.1), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1890, label = "N3s-OsHH", group = """ 1 * N3s u0 {2,S} {3,S} {4,S} 2 O2s u0 {1,S} 3 H u0 {1,S} 4 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([6.1,7.4,8.4,9.3,10.5,11.5,13.2],'cal/(mol*K)'), H298 = (11.4,'kcal/mol'), S298 = (29.1,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1904, label = "N3d", group = """ 1 * N3d u0 """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1906, label = "N3d-CdH", group = """ 1 * N3d u0 {2,D} {3,S} 2 [Cd,Cdd] u0 {1,D} 3 H u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3,3.5,3.9,4.3,5,5.5,6.4],'cal/(mol*K)'), H298 = (16.3,'kcal/mol'), S298 = (13.3,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1815, label = "N3d-N3dH", group = """ 1 * N3d u0 {2,S} {3,D} 2 H u0 {1,S} 3 N3d u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([4.38,4.89,5.44,5.94,6.77,7.42,8.44],'cal/(mol*K)'), H298 = (25.1,'kcal/mol'), S298 = (26.8,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1822, label = "N3d-N3dN3s", group = """ 1 * N3d u0 {2,D} {3,S} 2 N3d u0 {1,D} 3 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (23,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1909, label = "N3d-OdOs", group = """ 1 * N3d u0 {2,D} {3,S} 2 O2d u0 {1,D} 3 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1910, label = "N3d-OdN3s", group = """ 1 * N3d u0 {2,D} {3,S} 2 O2d u0 {1,D} 3 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1911, label = "N3d-CsR", group = """ 1 * N3d u0 {2,S} {3,D} 2 Cs u0 {1,S} 3 R!H u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (21.3,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1908, label = "N3d-OdC", group = """ 1 * N3d u0 {2,D} {3,S} 2 O2d u0 {1,D} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1905, label = "N3d-CdCs", group = """ 1 * N3d u0 {2,D} {3,S} 2 [Cd,Cdd] u0 {1,D} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([2,2.2,2.2,2.3,2.5,2.7,2.9],'cal/(mol*K)'), H298 = (21.3,'kcal/mol'), S298 = (-6.3,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1907, label = "N3d-N3dCs", group = """ 1 * N3d u0 {2,D} {3,S} 2 N3d u0 {1,D} 3 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([3.4,3.6,3.7,3.9,4.3,4.6,4.9],'cal/(mol*K)'), H298 = (27,'kcal/mol'), S298 = (7.2,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1912, label = "N3d-CbR", group = """ 1 * N3d u0 {2,S} {3,D} 2 Cb u0 {1,S} 3 R!H u0 {1,D} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (16.7,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1913, label = "N5dc", group = """ 1 * N5dc u0 """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1914, label = "N5dc-OdOsCs", group = """ 1 * N5dc u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 O2s u0 {1,S} 4 Cs u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1915, label = "N5dc-OdOsCd", group = """ 1 * N5dc u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 O2s u0 {1,S} 4 Cd u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1916, label = "N5dc-OdOsOs", group = """ 1 * N5dc u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 O2s u0 {1,S} 4 O2s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1917, label = "N5dc-OdOsN3s", group = """ 1 * N5dc u0 {2,D} {3,S} {4,S} 2 O2d u0 {1,D} 3 O2s u0 {1,S} 4 N3s u0 {1,S} """, thermo = ThermoData( Tdata = ([300,400,500,600,800,1000,1500],'K'), Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'), H298 = (0,'kcal/mol'), S298 = (0,'cal/(mol*K)'), ), shortDesc = u"""""", longDesc = u""" """, ) entry( index = 1918, label = "N5ddc", group = """ 1 * N5ddc u0 """, thermo = None, shortDesc = u"""""", longDesc = u""" """, ) tree( """ L1: R L2: C L3: Cbf L4: Cbf-CbCbCbf L4: Cbf-CbCbfCbf L4: Cbf-CbfCbfCbf L3: Cb L4: Cb-H L4: Cb-O2s L4: Cb-S2s L4: Cb-N3s L4: Cb-C L5: Cb-Cs L5: Cb-Cds L6: Cb-(Cds-O2d) L6: Cb-(Cds-Cd) L7: Cb-(Cds-Cds) L7: Cb-(Cds-Cdd) L8: Cb-(Cds-Cdd-O2d) L8: Cb-(Cds-Cdd-S2d) L8: Cb-(Cds-Cdd-Cd) L5: Cb-Ct L6: Cb-(CtN3t) L5: Cb-Cb L5: Cb-C=S L3: Ct L4: Ct-CtN3s L4: Ct-N3tN3s L4: Ct-CtH L4: Ct-CtOs L4: Ct-N3tOs L4: Ct-CtSs L4: Ct-N3tC L5: Ct-N3tCs L5: Ct-N3tCd L4: Ct-CtC L5: Ct-CtCs L5: Ct-CtCds L6: Ct-Ct(Cds-O2d) L6: Ct-Ct(Cds-Cd) L7: Ct-Ct(Cds-Cds) L7: Ct-Ct(Cds-Cdd) L8: Ct-Ct(Cds-Cdd-O2d) L8: Ct-Ct(Cds-Cdd-S2d) L8: Ct-Ct(Cds-Cdd-Cd) L5: Ct-CtCt L6: Ct-Ct(CtN3t) L5: Ct-CtCb L5: Ct-CtC=S L3: Cdd L4: Cdd-N3dCd L4: Cdd-OdOd L4: Cdd-OdSd L4: Cdd-SdSd L4: Cdd-CdOd L5: Cdd-CdsOd L5: Cdd-CddOd L6: Cdd-(Cdd-O2d)O2d L6: Cdd-(Cdd-Cd)O2d L4: Cdd-CdSd L5: Cdd-CdsSd L5: Cdd-CddSd L6: Cdd-(Cdd-S2d)S2d L6: Cdd-(Cdd-Cd)S2d L4: Cdd-CdCd L5: Cdd-CddCdd L6: Cdd-(Cdd-O2d)(Cdd-O2d) L6: Cdd-(Cdd-S2d)(Cdd-S2d) L6: Cdd-(Cdd-O2d)(Cdd-Cd) L6: Cdd-(Cdd-S2d)(Cdd-Cd) L6: Cdd-(Cdd-Cd)(Cdd-Cd) L5: Cdd-CddCds L6: Cdd-(Cdd-O2d)Cds L6: Cdd-(Cdd-S2d)Cds L6: Cdd-(Cdd-Cd)Cds L5: Cdd-CdsCds L3: Cds L4: Cds-OdN3sH L4: Cds-OdN3sCs L4: Cd-N3dCsCs L4: Cd-N3dCsH L4: Cd-N3dHH L4: Cds-OdHH L4: Cds-OdOsH L4: CO-SsH L4: Cds-OdOsOs L4: CO-CsSs L4: CO-OsSs L4: Cds-OdCH L5: Cds-OdCsH L5: Cds-OdCdsH L6: Cds-O2d(Cds-O2d)H L6: Cds-O2d(Cds-Cd)H L7: Cds-O2d(Cds-Cds)H L7: Cds-O2d(Cds-Cdd)H L8: Cds-O2d(Cds-Cdd-O2d)H L8: Cds-O2d(Cds-Cdd-Cd)H L5: Cds-OdCtH L5: Cds-OdCbH L4: Cds-OdCOs L5: Cds-OdCsOs L5: Cds-OdCdsOs L6: Cds-O2d(Cds-O2d)O2s L6: Cds-O2d(Cds-Cd)O2s L7: Cds-O2d(Cds-Cds)O2s L7: Cds-O2d(Cds-Cdd)O2s L8: Cds-O2d(Cds-Cdd-O2d)O2s L8: Cds-O2d(Cds-Cdd-Cd)O2s L5: Cds-OdCtOs L5: Cds-OdCbOs L4: Cds-OdCC L5: Cds-OdCsCs L5: Cds-OdCdsCs L6: Cds-O2d(Cds-O2d)Cs L6: Cds-O2d(Cds-Cd)Cs L7: Cds-O2d(Cds-Cds)Cs L7: Cds-O2d(Cds-Cdd)Cs L8: Cds-O2d(Cds-Cdd-O2d)Cs L8: Cds-O2d(Cds-Cdd-Cd)Cs L5: Cds-OdCdsCds L6: Cds-O2d(Cds-O2d)(Cds-O2d) L6: Cds-O2d(Cds-Cd)(Cds-O2d) L7: Cds-O2d(Cds-Cds)(Cds-O2d) L7: Cds-O2d(Cds-Cdd)(Cds-O2d) L8: Cds-O2d(Cds-Cdd-O2d)(Cds-O2d) L8: Cds-O2d(Cds-Cdd-Cd)(Cds-O2d) L6: Cds-O2d(Cds-Cd)(Cds-Cd) L7: Cds-O2d(Cds-Cds)(Cds-Cds) L7: Cds-O2d(Cds-Cdd)(Cds-Cds) L8: Cds-O2d(Cds-Cdd-O2d)(Cds-Cds) L8: Cds-O2d(Cds-Cdd-Cd)(Cds-Cds) L7: Cds-O2d(Cds-Cdd)(Cds-Cdd) L8: Cds-O2d(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-O2d) L8: Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-Cd) L5: Cds-OdCtCs L5: Cds-OdCtCds L6: Cds-OdCt(Cds-O2d) L6: Cds-OdCt(Cds-Cd) L7: Cds-OdCt(Cds-Cds) L7: Cds-OdCt(Cds-Cdd) L8: Cds-OdCt(Cds-Cdd-O2d) L8: Cds-OdCt(Cds-Cdd-Cd) L5: Cds-OdCtCt L5: Cds-OdCbCs L5: Cds-OdCbCds L6: Cds-OdCb(Cds-O2d) L6: Cds-OdCb(Cds-Cd) L7: Cds-OdCb(Cds-Cds) L7: Cds-OdCb(Cds-Cdd) L8: Cds-OdCb(Cds-Cdd-O2d) L8: Cds-OdCb(Cds-Cdd-Cd) L5: Cds-OdCbCt L5: Cds-OdCbCb L4: Cds-CdHH L5: Cds-CdsHH L5: Cds-CddHH L6: Cds-(Cdd-O2d)HH L6: Cds-(Cdd-S2d)HH L6: Cds-(Cdd-Cd)HH L4: Cds-CdOsH L5: Cds-CdsOsH L5: Cds-CddOsH L6: Cds-(Cdd-O2d)OsH L6: Cds-(Cdd-Cd)OsH L4: Cds-CdSsH L5: Cds-CdsSsH L5: Cds-CddSsH L6: Cds-(Cdd-S2d)SsH L6: Cds-(Cdd-Cd)SsH L4: Cds-CdOsOs L5: Cds-CdsOsOs L5: Cds-CddOsOs L6: Cds-(Cdd-O2d)OsOs L6: Cds-(Cdd-Cd)OsOs L4: Cds-CdSsSs L5: Cds-CdsSsSs L5: Cds-CddSsSs L6: Cds-(Cdd-S2d)SsSs L6: Cds-(Cdd-Cd)SsSs L4: Cds-CdCH L5: Cds-CdsCsH L5: Cds-CdsCdsH L6: Cd-Cd(CO)H L6: Cds-Cds(Cds-Cd)H L7: Cds-Cds(Cds-Cds)H L7: Cds-Cds(Cds-Cdd)H L8: Cd-Cd(CCO)H L8: Cds-Cds(Cds-Cdd-S2d)H L8: Cds-Cds(Cds-Cdd-Cd)H L5: Cds-CdsCtH L6: Cds-CdsH(CtN3t) L5: Cds-CdsCbH L5: Cds-CddCsH L6: Cds-(Cdd-O2d)CsH L6: Cds-(Cdd-S2d)CsH L6: Cds-(Cdd-Cd)CsH L5: Cds-CddCdsH L6: Cds-(Cdd-O2d)(Cds-O2d)H L6: Cds-(Cdd-O2d)(Cds-Cd)H L7: Cds-(Cdd-O2d)(Cds-Cds)H L7: Cds-(Cdd-O2d)(Cds-Cdd)H L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)H L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)H L6: Cds-(Cdd-S2d)(Cds-Cd)H L7: Cds-(Cdd-S2d)(Cds-Cds)H L7: Cds-(Cdd-S2d)(Cds-Cdd)H L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)H L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)H L6: Cds-(Cdd-Cd)(Cds-O2d)H L6: Cds-(Cdd-Cd)(Cds-Cd)H L7: Cds-(Cdd-Cd)(Cds-Cds)H L7: Cds-(Cdd-Cd)(Cds-Cdd)H L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)H L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)H L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)H L5: Cds-CddCtH L6: Cds-(Cdd-O2d)CtH L6: Cds-(Cdd-S2d)CtH L6: Cds-(Cdd-Cd)CtH L5: Cds-CddCbH L6: Cds-(Cdd-O2d)CbH L6: Cds-(Cdd-S2d)CbH L6: Cds-(Cdd-Cd)CbH L5: Cds-(Cdd-Cd)C=SH L5: Cds-(Cdd-S2d)C=SH L5: Cds-CdsC=SH L4: Cds-CdCO L5: Cds-CdsCdsOs L6: Cds-Cds(Cds-O2d)O2s L6: Cds-Cds(Cds-Cd)O2s L7: Cds-Cds(Cds-Cds)O2s L7: Cds-Cds(Cds-Cdd)O2s L8: Cds-Cds(Cds-Cdd-O2d)O2s L8: Cds-Cds(Cds-Cdd-Cd)O2s L5: Cds-CdsCtOs L5: Cds-CdsCbOs L5: Cds-CddCdsOs L6: Cds-(Cdd-O2d)(Cds-O2d)O2s L6: Cds-(Cdd-O2d)(Cds-Cd)O2s L7: Cds-(Cdd-O2d)(Cds-Cds)O2s L7: Cds-(Cdd-O2d)(Cds-Cdd)O2s L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)O2s L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)O2s L6: Cds-(Cdd-Cd)(Cds-Cd)O2s L7: Cds-(Cdd-Cd)(Cds-Cds)O2s L7: Cds-(Cdd-Cd)(Cds-Cdd)O2s L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)O2s L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)O2s L5: Cds-CddCtOs L6: Cds-(Cdd-O2d)CtOs L6: Cds-(Cdd-Cd)CtOs L5: Cds-CddCbOs L6: Cds-(Cdd-O2d)CbOs L6: Cds-(Cdd-Cd)CbOs L5: Cd-CdCsOs L6: Cds-CdsCsOs L6: Cds-CddCsOs L7: Cds-(Cdd-O2d)CsOs L7: Cds-(Cdd-Cd)CsOs L4: Cds-CdCS L5: Cds-CdsCsSs L5: Cds-CdsCdsSs L6: Cds-Cds(Cds-Cd)S2s L7: Cds-Cds(Cds-Cds)S2s L7: Cds-Cds(Cds-Cdd)S2s L8: Cds-Cds(Cds-Cdd-S2d)S2s L8: Cds-Cds(Cds-Cdd-Cd)S2s L5: Cds-CdsCtSs L5: Cds-CdsCbSs L5: Cds-CddCsSs L6: Cds-(Cdd-S2d)CsSs L6: Cds-(Cdd-Cd)CsSs L5: Cds-CddCdsSs L6: Cds-(Cdd-S2d)(Cds-Cd)S2s L7: Cds-(Cdd-S2d)(Cds-Cds)S2s L7: Cds-(Cdd-S2d)(Cds-Cdd)S2s L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)S2s L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)S2s L6: Cds-(Cdd-Cd)(Cds-Cd)S2s L7: Cds-(Cdd-Cd)(Cds-Cds)S2s L7: Cds-(Cdd-Cd)(Cds-Cdd)S2s L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)S2s L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)S2s L5: Cds-CddCtSs L6: Cds-(Cdd-S2d)CtSs L6: Cds-(Cdd-Cd)CtSs L5: Cds-CddCbSs L6: Cds-(Cdd-S2d)CbSs L6: Cds-(Cdd-Cd)CbSs L5: Cds-(Cdd-S2d)C=SSs L5: Cds-CdsC=SSs L4: Cds-CdCC L5: Cds-CdsCsCs L5: Cds-CdsCdsCs L6: Cd-CdCs(CO) L6: Cds-Cds(Cds-Cd)Cs L7: Cds-Cds(Cds-Cds)Cs L7: Cds-Cds(Cds-Cdd)Cs L8: Cd-CdCs(CCO) L8: Cds-Cds(Cds-Cdd-S2d)Cs L8: Cds-Cds(Cds-Cdd-Cd)Cs L5: Cds-CdsCdsCds L6: Cds-Cds(Cds-O2d)(Cds-O2d) L6: Cds-Cds(Cds-O2d)(Cds-Cd) L7: Cds-Cds(Cds-O2d)(Cds-Cds) L7: Cds-Cds(Cds-O2d)(Cds-Cdd) L8: Cds-Cds(Cds-O2d)(Cds-Cdd-O2d) L8: Cds-Cds(Cds-O2d)(Cds-Cdd-Cd) L6: Cds-Cds(Cds-Cd)(Cds-Cd) L7: Cds-Cds(Cds-Cds)(Cds-Cds) L7: Cds-Cds(Cds-Cds)(Cds-Cdd) L8: Cds-Cds(Cds-Cds)(Cds-Cdd-O2d) L8: Cds-Cds(Cds-Cds)(Cds-Cdd-S2d) L8: Cds-Cds(Cds-Cds)(Cds-Cdd-Cd) L7: Cds-Cds(Cds-Cdd)(Cds-Cdd) L8: Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cds-Cds(Cds-Cdd-S2d)(Cds-Cdd-S2d) L8: Cds-Cds(Cds-Cdd-S2d)(Cds-Cdd-Cd) L8: Cds-Cds(Cds-Cdd-Cd)(Cds-Cdd-Cd) L5: Cds-CdsCtCs L6: Cd-CdCs(CtN3t) L5: Cds-CdsCtCds L6: Cds-CdsCt(Cds-O2d) L6: Cds-CdsCt(Cds-Cd) L7: Cds-Cds(Cds-Cds)Ct L7: Cds-Cds(Cds-Cdd)Ct L8: Cds-Cds(Cds-Cdd-O2d)Ct L8: Cds-Cds(Cds-Cdd-S2d)Ct L8: Cds-Cds(Cds-Cdd-Cd)Ct L5: Cds-CdsCtCt L6: Cds-Cd(CtN3t)(CtN3t) L5: Cds-CdsCbCs L5: Cds-CdsCbCds L6: Cds-CdsCb(Cds-O2d) L6: Cds-Cds(Cds-Cd)Cb L7: Cds-Cds(Cds-Cds)Cb L7: Cds-Cds(Cds-Cdd)Cb L8: Cds-Cds(Cds-Cdd-O2d)Cb L8: Cds-Cds(Cds-Cdd-S2d)Cb L8: Cds-Cds(Cds-Cdd-Cd)Cb L5: Cds-CdsCbCt L5: Cds-CdsCbCb L5: Cds-CddCsCs L6: Cds-(Cdd-O2d)CsCs L6: Cds-(Cdd-S2d)CsCs L6: Cds-(Cdd-Cd)CsCs L5: Cds-CddCdsCs L6: Cds-(Cdd-O2d)(Cds-O2d)Cs L6: Cds-(Cdd-O2d)(Cds-Cd)Cs L7: Cds-(Cdd-O2d)(Cds-Cds)Cs L7: Cds-(Cdd-O2d)(Cds-Cdd)Cs L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cs L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cs L6: Cds-(Cdd-S2d)(Cds-Cd)Cs L7: Cds-(Cdd-S2d)(Cds-Cds)Cs L7: Cds-(Cdd-S2d)(Cds-Cdd)Cs L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)Cs L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)Cs L6: Cds-(Cdd-Cd)(Cds-Cd)Cs L7: Cds-(Cdd-Cd)(Cds-Cds)Cs L7: Cds-(Cdd-Cd)(Cds-Cdd)Cs L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)Cs L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)Cs L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cs L5: Cds-CddCdsCds L6: Cds-(Cdd-O2d)(Cds-O2d)(Cds-O2d) L6: Cds-(Cdd-O2d)(Cds-Cd)(Cds-O2d) L7: Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d) L7: Cds-(Cdd-O2d)(Cds-Cdd)(Cds-O2d) L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-O2d) L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-O2d) L6: Cds-(Cdd-O2d)(Cds-Cd)(Cds-Cd) L7: Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds) L7: Cds-(Cdd-O2d)(Cds-Cdd)(Cds-Cds) L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds) L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cds) L7: Cds-(Cdd-O2d)(Cds-Cdd)(Cds-Cdd) L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L6: Cds-(Cdd-Cd)(Cds-O2d)(Cds-O2d) L6: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cd) L7: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cds) L7: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd) L8: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-O2d) L8: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-Cd) L6: Cds-(Cdd-S2d)(Cds-Cd)(Cds-Cd) L7: Cds-(Cdd-S2d)(Cds-Cds)(Cds-Cds) L7: Cds-(Cdd-S2d)(Cds-Cdd)(Cds-Cds) L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cds) L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cds) L7: Cds-(Cdd-S2d)(Cds-Cdd)(Cds-Cdd) L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d) L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd) L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L6: Cds-(Cdd-Cd)(Cds-Cd)(Cds-Cd) L7: Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds) L7: Cds-(Cdd-Cd)(Cds-Cdd)(Cds-Cds) L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cds) L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cds) L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cds) L7: Cds-(Cdd-Cd)(Cds-Cdd)(Cds-Cdd) L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cdd-S2d) L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cdd-Cd) L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L5: Cds-CddCtCs L6: Cds-(Cdd-O2d)CtCs L6: Cds-(Cdd-S2d)CtCs L6: Cds-(Cdd-Cd)CtCs L5: Cds-CddCtCds L6: Cds-(Cdd-O2d)(Cds-O2d)Ct L6: Cds-(Cdd-O2d)(Cds-Cd)Ct L7: Cds-(Cdd-O2d)(Cds-Cds)Ct L7: Cds-(Cdd-O2d)(Cds-Cdd)Ct L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)Ct L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)Ct L6: Cds-(Cdd-S2d)(Cds-Cd)Ct L7: Cds-(Cdd-S2d)(Cds-Cds)Ct L7: Cds-(Cdd-S2d)(Cds-Cdd)Ct L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)Ct L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)Ct L6: Cds-(Cdd-Cd)(Cds-Cd)Ct L7: Cds-(Cdd-Cd)(Cds-Cds)Ct L7: Cds-(Cdd-Cd)(Cds-Cdd)Ct L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)Ct L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)Ct L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)Ct L5: Cds-CddCtCt L6: Cds-(Cdd-O2d)CtCt L6: Cds-(Cdd-S2d)CtCt L6: Cds-(Cdd-Cd)CtCt L5: Cds-CddCbCs L6: Cds-(Cdd-O2d)CbCs L6: Cds-(Cdd-S2d)CbCs L6: Cds-(Cdd-Cd)CbCs L5: Cds-CddCbCds L6: Cds-(Cdd-O2d)(Cds-O2d)Cb L6: Cds-(Cdd-O2d)(Cds-Cd)Cb L7: Cds-(Cdd-O2d)(Cds-Cds)Cb L7: Cds-(Cdd-O2d)(Cds-Cdd)Cb L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cb L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cb L6: Cds-(Cdd-S2d)(Cds-Cd)Cb L7: Cds-(Cdd-S2d)(Cds-Cds)Cb L7: Cds-(Cdd-S2d)(Cds-Cdd)Cb L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)Cb L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)Cb L6: Cds-(Cdd-Cd)(Cds-Cd)Cb L7: Cds-(Cdd-Cd)(Cds-Cds)Cb L7: Cds-(Cdd-Cd)(Cds-Cdd)Cb L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)Cb L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)Cb L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cb L5: Cds-CddCbCt L6: Cds-(Cdd-O2d)CbCt L6: Cds-(Cdd-S2d)CbCt L6: Cds-(Cdd-Cd)CbCt L5: Cds-CddCbCb L6: Cds-(Cdd-O2d)CbCb L6: Cds-(Cdd-S2d)CbCb L6: Cds-(Cdd-Cd)CbCb L5: Cds-CdsC=SC=S L5: Cds-(Cdd-Cd)C=S(Cds-Cd) L6: Cds-(Cdd-Cd)C=S(Cds-Cds) L6: Cds-(Cdd-Cd)C=S(Cds-Cdd) L7: Cds-(Cdd-Cd)C=S(Cds-Cdd-Cd) L7: Cds-(Cdd-Cd)C=S(Cds-Cdd-S2d) L5: Cds-(Cdd-S2d)C=SCs L5: Cds-(Cdd-S2d)C=SCt L5: Cds-(Cdd-S2d)C=SCb L5: Cds-(Cdd-Cd)C=SC=S L5: Cds-(Cdd-S2d)(Cds-Cd)C=S L6: Cds-(Cdd-S2d)(Cds-Cds)C=S L6: Cds-(Cdd-S2d)(Cds-Cdd)C=S L7: Cds-(Cdd-S2d)(Cds-Cdd-S2d)C=S L7: Cds-(Cdd-S2d)(Cds-Cdd-Cd)C=S L5: Cds-CdsCbC=S L5: Cds-CdsCtC=S L5: Cds-CdsC=SCs L5: Cds-CdsC=S(Cds-Cd) L6: Cds-CdsC=S(Cds-Cds) L6: Cds-CdsC=S(Cds-Cdd) L7: Cds-CdsC=S(Cds-Cdd-Cd) L7: Cds-CdsC=S(Cds-Cdd-S2d) L5: Cds-(Cdd-S2d)C=SC=S L4: Cds-CNH L5: Cd-CdHN3s L5: Cd-CdH(N5dOdOs) L4: Cds-CCN L5: Cd-CdCsN3s L5: Cd-CdCs(N5dOdOs) L4: C=S-SsSs L4: C=S-CH L5: C=S-CsH L5: C=S-CdsH L6: C=S-(Cds-Cd)H L7: C=S-(Cds-Cdd)H L8: C=S-(Cds-Cdd-Cd)H L8: C=S-(Cds-Cdd-S2d)H L7: C=S-(Cds-Cds)H L5: C=S-CtH L5: C=S-CbH L5: C=S-C=SH L4: C=S-CC L5: C=S-CbCds L6: C=S-Cb(Cds-Cd) L7: C=S-Cb(Cds-Cds) L7: C=S-Cb(Cds-Cdd) L8: C=S-Cb(Cds-Cdd-S2d) L8: C=S-Cb(Cds-Cdd-Cd) L5: C=S-CtCt L5: C=S-CbCb L5: C=S-CdsCds L6: C=S-(Cds-Cd)(Cds-Cd) L7: C=S-(Cds-Cdd)(Cds-Cds) L8: C=S-(Cds-Cdd-Cd)(Cds-Cds) L8: C=S-(Cds-Cdd-S2d)(Cds-Cds) L7: C=S-(Cds-Cds)(Cds-Cds) L7: C=S-(Cds-Cdd)(Cds-Cdd) L8: C=S-(Cds-Cdd-Cd)(Cds-Cdd-Cd) L8: C=S-(Cds-Cdd-S2d)(Cds-Cdd-S2d) L8: C=S-(Cds-Cdd-Cd)(Cds-Cdd-S2d) L5: C=S-CtCds L6: C=S-Ct(Cds-Cd) L7: C=S-Ct(Cds-Cds) L7: C=S-Ct(Cds-Cdd) L8: C=S-Ct(Cds-Cdd-Cd) L8: C=S-Ct(Cds-Cdd-S2d) L5: C=S-CbCt L5: C=S-CsCs L5: C=S-CdsCs L6: C=S-(Cds-Cd)Cs L7: C=S-(Cds-Cds)Cs L7: C=S-(Cds-Cdd)Cs L8: C=S-(Cds-Cdd-S2d)Cs L8: C=S-(Cds-Cdd-Cd)Cs L5: C=S-CtCs L5: C=S-CbCs L5: C=S-C=SCs L5: C=S-CtC=S L5: C=S-(Cds-Cd)C=S L6: C=S-(Cds-Cdd)C=S L7: C=S-(Cds-Cdd-Cd)C=S L7: C=S-(Cds-Cdd-S2d)C=S L6: C=S-(Cds-Cds)C=S L5: C=S-C=SC=S L5: C=S-CbC=S L4: C=S-HH L4: C=S-SsH L4: C=S-CSs L5: C=S-CbSs L5: C=S-CdsSs L6: C=S-(Cds-Cd)S2s L7: C=S-(Cds-Cds)S2s L7: C=S-(Cds-Cdd)S2s L8: C=S-(Cds-Cdd-Cd)S2s L8: C=S-(Cds-Cdd-S2d)S2s L5: C=S-CtSs L5: C=S-CsSs L5: C=S-C=SSs L4: CS-OsH L4: CS-CsOs L4: CS-OsOs L3: Cs L4: Cs-NHHH L5: Cs-N3sHHH L5: Cs-N3dHHH L6: Cs-(N3dCd)HHH L6: Cs-(N3dN3d)HHH L4: Cs-NCsHH L5: Cs-N3sCsHH L5: Cs-N3dCHH L6: Cs-(N3dN3d)CsHH L6: Cs-(N3dOd)CHH L6: Cs-(N3dCd)CsHH L5: Cs-N5dCsHH L6: Cs-(N5dOdOs)CsHH L4: Cs-NCsCsH L5: Cs-N3sCsCsH L5: Cs-N3dCsCsH L6: Cs-(N3dOd)CsCsH L6: Cs-(N3dN3d)CsCsH L5: Cs-N5dCsCsH L6: Cs-(N5dOdOs)CsCsH L4: Cs-NCsCsCs L5: Cs-N3sCsCsCs L5: Cs-N3dCsCsCs L6: Cs-(N3dOd)CsCsCs L6: Cs-(N3dN3d)CsCsCs L5: Cs-N5dCsCsCs L6: Cs-(N5dOdOs)CsCsCs L4: Cs-NNCsCs L5: Cs-N5dN5dCsCs L4: Cs-NNCsH L5: Cs-(N5dOdOs)(N5dOdOs)CsH L4: Cs-HHHH L4: Cs-CHHH L5: Cs-CsHHH L5: Cs-CdsHHH L6: Cs-(Cds-O2d)HHH L6: Cs-(Cds-Cd)HHH L7: Cs-(Cds-Cds)HHH L7: Cs-(Cds-Cdd)HHH L8: Cs-(Cds-Cdd-O2d)HHH L8: Cs-(Cds-Cdd-S2d)HHH L8: Cs-(Cds-Cdd-Cd)HHH L6: Cs-(CdN3d)HHH L5: Cs-CtHHH L6: Cs-(CtN3t)HHH L5: Cs-CbHHH L5: Cs-C=SHHH L4: Cs-OsHHH L4: Cs-OsOsHH L4: Cs-OsOsOsH L4: Cs-OsSsHH L4: Cs-OsOsSsH L4: Cs-SsHHH L4: Cs-SsSsHH L4: Cs-SsSsSsH L4: Cs-CCHH L5: Cs-CsCsHH L5: Cs-CdsCsHH L6: Cs-(Cds-O2d)CsHH L6: Cs-(Cds-Cd)CsHH L7: Cs-(Cds-Cds)CsHH L7: Cs-(Cds-Cdd)CsHH L8: Cs-(Cds-Cdd-O2d)CsHH L8: Cs-(Cds-Cdd-S2d)CsHH L8: Cs-(Cds-Cdd-Cd)CsHH L6: Cs-(CdN3d)CsHH L5: Cs-CdsCdsHH L6: Cs-(Cds-O2d)(Cds-O2d)HH L6: Cs-(Cds-O2d)(Cds-Cd)HH L7: Cs-(Cds-O2d)(Cds-Cds)HH L7: Cs-(Cds-O2d)(Cds-Cdd)HH L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)HH L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)HH L6: Cs-(Cds-Cd)(Cds-Cd)HH L7: Cs-(Cds-Cds)(Cds-Cds)HH L7: Cs-(Cds-Cdd)(Cds-Cds)HH L8: Cs-Cd(CCO)HH L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)HH L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)HH L7: Cs-(Cds-Cdd)(Cds-Cdd)HH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)HH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)HH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)HH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)HH L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)HH L5: Cs-CtCsHH L6: Cs-(CtN3t)CsHH L5: Cs-CtCdsHH L6: Cs-(Cds-O2d)CtHH L6: Cs-(Cds-Cd)CtHH L7: Cs-(Cds-Cds)CtHH L7: Cs-(Cds-Cdd)CtHH L8: Cs-(Cds-Cdd-O2d)CtHH L8: Cs-(Cds-Cdd-S2d)CtHH L8: Cs-(Cds-Cdd-Cd)CtHH L5: Cs-CtCtHH L5: Cs-CbCsHH L5: Cs-CbCdsHH L6: Cs-(Cds-O2d)CbHH L6: Cs-(Cds-Cd)CbHH L7: Cs-(Cds-Cds)CbHH L7: Cs-(Cds-Cdd)CbHH L8: Cs-(Cds-Cdd-O2d)CbHH L8: Cs-(Cds-Cdd-S2d)CbHH L8: Cs-(Cds-Cdd-Cd)CbHH L5: Cs-CbCtHH L5: Cs-CbCbHH L5: Cs-C=SCtHH L5: Cs-C=SCsHH L5: Cs-C=S(Cds-Cd)HH L6: Cs-C=S(Cds-Cdd)HH L7: Cs-C=S(Cds-Cdd-Cd)HH L7: Cs-C=S(Cds-Cdd-S2d)HH L6: Cs-C=S(Cds-Cds)HH L5: Cs-C=SC=SHH L5: Cs-C=SCbHH L4: Cs-CCCH L5: Cs-CsCsCsH L5: Cs-CdsCsCsH L6: Cs-(Cds-O2d)CsCsH L6: Cs-(Cds-Cd)CsCsH L7: Cs-(Cds-Cds)CsCsH L7: Cs-(Cds-Cdd)CsCsH L8: Cs-(Cds-Cdd-O2d)CsCsH L8: Cs-(Cds-Cdd-S2d)CsCsH L8: Cs-(Cds-Cdd-Cd)CsCsH L6: Cs-(CdN3d)CsCsH L5: Cs-CtCsCsH L6: Cs-(CtN3t)CsCsH L5: Cs-CbCsCsH L5: Cs-CdsCdsCsH L6: Cs-(Cds-O2d)(Cds-O2d)CsH L6: Cs-(Cds-O2d)(Cds-Cd)CsH L7: Cs-(Cds-O2d)(Cds-Cds)CsH L7: Cs-(Cds-O2d)(Cds-Cdd)CsH L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CsH L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CsH L6: Cs-(Cds-Cd)(Cds-Cd)CsH L7: Cs-(Cds-Cds)(Cds-Cds)CsH L7: Cs-(Cds-Cdd)(Cds-Cds)CsH L8: Cs-CsCd(CCO)H L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CsH L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CsH L7: Cs-(Cds-Cdd)(Cds-Cdd)CsH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsH L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsH L5: Cs-CtCdsCsH L6: Cs-(Cds-O2d)CtCsH L6: Cs-(Cds-Cd)CtCsH L7: Cs-(Cds-Cds)CtCsH L7: Cs-(Cds-Cdd)CtCsH L8: Cs-(Cds-Cdd-O2d)CtCsH L8: Cs-(Cds-Cdd-S2d)CtCsH L8: Cs-(Cds-Cdd-Cd)CtCsH L5: Cs-CbCdsCsH L6: Cs-(Cds-O2d)CbCsH L6: Cs-(Cds-Cd)CbCsH L7: Cs-(Cds-Cds)CbCsH L7: Cs-(Cds-Cdd)CbCsH L8: Cs-(Cds-Cdd-O2d)CbCsH L8: Cs-(Cds-Cdd-Cd)CbCsH L5: Cs-CtCtCsH L5: Cs-CbCtCsH L5: Cs-CbCbCsH L5: Cs-CdsCdsCdsH L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)H L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)H L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)H L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)H L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)H L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)H L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)H L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)H L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)H L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)H L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)H L8: Cs-CdCd(CCO)H L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)H L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)H L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)H L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)H L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)H L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)H L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)H L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)H L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H L5: Cs-CtCdsCdsH L6: Cs-(Cds-O2d)(Cds-O2d)CtH L6: Cs-(Cds-O2d)(Cds-Cd)CtH L7: Cs-(Cds-O2d)(Cds-Cds)CtH L7: Cs-(Cds-O2d)(Cds-Cdd)CtH L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CtH L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CtH L6: Cs-(Cds-Cd)(Cds-Cd)CtH L7: Cs-(Cds-Cds)(Cds-Cds)CtH L7: Cs-(Cds-Cdd)(Cds-Cds)CtH L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CtH L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CtH L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtH L7: Cs-(Cds-Cdd)(Cds-Cdd)CtH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtH L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtH L5: Cs-CbCdsCdsH L6: Cs-(Cds-O2d)(Cds-O2d)CbH L6: Cs-(Cds-O2d)(Cds-Cd)CbH L7: Cs-(Cds-O2d)(Cds-Cds)CbH L7: Cs-(Cds-O2d)(Cds-Cdd)CbH L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbH L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbH L6: Cs-(Cds-Cd)(Cds-Cd)CbH L7: Cs-(Cds-Cds)(Cds-Cds)CbH L7: Cs-(Cds-Cdd)(Cds-Cds)CbH L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbH L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbH L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbH L7: Cs-(Cds-Cdd)(Cds-Cdd)CbH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbH L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbH L5: Cs-CtCtCdsH L6: Cs-CtCt(Cds-O2d)H L6: Cs-CtCt(Cds-Cd)H L7: Cs-CtCt(Cds-Cds)H L7: Cs-CtCt(Cds-Cdd)H L8: Cs-CtCt(Cds-Cdd-O2d)H L8: Cs-CtCt(Cds-Cdd-S2d)H L8: Cs-CtCt(Cds-Cdd-Cd)H L5: Cs-CbCtCdsH L6: Cs-CbCt(Cds-O2d)H L6: Cs-CbCt(Cds-Cd)H L7: Cs-CbCt(Cds-Cds)H L7: Cs-CbCt(Cds-Cdd)H L8: Cs-CbCt(Cds-Cdd-O2d)H L8: Cs-CbCt(Cds-Cdd-S2d)H L8: Cs-CbCt(Cds-Cdd-Cd)H L5: Cs-CbCbCdsH L6: Cs-CbCb(Cds-O2d)H L6: Cs-CbCb(Cds-Cd)H L7: Cs-CbCb(Cds-Cds)H L7: Cs-CbCb(Cds-Cdd)H L8: Cs-CbCb(Cds-Cdd-O2d)H L8: Cs-CbCb(Cds-Cdd-S2d)H L8: Cs-CbCb(Cds-Cdd-Cd)H L5: Cs-CtCtCtH L5: Cs-CbCtCtH L5: Cs-CbCbCtH L5: Cs-CbCbCbH L5: Cs-C=SC=SCbH L5: Cs-C=S(Cds-Cd)(Cds-Cd)H L6: Cs-C=S(Cds-Cdd)(Cds-Cds)H L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)H L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)H L6: Cs-C=S(Cds-Cds)(Cds-Cds)H L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)H L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)H L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)H L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)H L5: Cs-C=S(Cds-Cd)CtH L6: Cs-C=S(Cds-Cdd)CtH L7: Cs-C=S(Cds-Cdd-S2d)CtH L7: Cs-C=S(Cds-Cdd-Cd)CtH L6: Cs-C=S(Cds-Cds)CtH L5: Cs-C=SC=SCtH L5: Cs-C=SCtCsH L5: Cs-C=SC=SCsH L5: Cs-C=S(Cds-Cd)CbH L6: Cs-C=S(Cds-Cds)CbH L6: Cs-C=S(Cds-Cdd)CbH L7: Cs-C=S(Cds-Cdd-S2d)CbH L7: Cs-C=S(Cds-Cdd-Cd)CbH L5: Cs-C=S(Cds-Cd)CsH L6: Cs-C=S(Cds-Cds)CsH L6: Cs-C=S(Cds-Cdd)CsH L7: Cs-C=S(Cds-Cdd-Cd)CsH L7: Cs-C=S(Cds-Cdd-S2d)CsH L5: Cs-CbCtC=SH L5: Cs-C=SC=SC=SH L5: Cs-C=SCsCsH L5: Cs-CtCtC=SH L5: Cs-CbCbC=SH L5: Cs-C=SC=S(Cds-Cd)H L6: Cs-C=SC=S(Cds-Cds)H L6: Cs-C=SC=S(Cds-Cdd)H L7: Cs-C=SC=S(Cds-Cdd-S2d)H L7: Cs-C=SC=S(Cds-Cdd-Cd)H L4: Cs-CCCC L5: Cs-CsCsCsCs L5: Cs-CdsCsCsCs L6: Cs-(Cds-O2d)CsCsCs L6: Cs-(Cds-Cd)CsCsCs L7: Cs-(Cds-Cds)CsCsCs L7: Cs-(Cds-Cdd)CsCsCs L8: Cs-(Cds-Cdd-O2d)CsCsCs L8: Cs-(Cds-Cdd-S2d)CsCsCs L8: Cs-(Cds-Cdd-Cd)CsCsCs L6: Cs-(CdN3d)CsCsCs L5: Cs-CtCsCsCs L6: Cs-(CtN3t)CsCsCs L5: Cs-CbCsCsCs L5: Cs-CdsCdsCsCs L6: Cs-(Cds-O2d)(Cds-O2d)CsCs L6: Cs-(Cds-O2d)(Cds-Cd)CsCs L7: Cs-(Cds-O2d)(Cds-Cds)CsCs L7: Cs-(Cds-O2d)(Cds-Cdd)CsCs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CsCs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CsCs L6: Cs-(Cds-Cd)(Cds-Cd)CsCs L7: Cs-(Cds-Cds)(Cds-Cds)CsCs L7: Cs-(Cds-Cdd)(Cds-Cds)CsCs L8: Cs-CsCsCd(CCO) L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CsCs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CsCs L7: Cs-(Cds-Cdd)(Cds-Cdd)CsCs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsCs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsCs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsCs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsCs L5: Cs-CtCdsCsCs L6: Cs-(Cds-O2d)CtCsCs L6: Cs-(Cds-Cd)CtCsCs L7: Cs-(Cds-Cds)CtCsCs L7: Cs-(Cds-Cdd)CtCsCs L8: Cs-(Cds-Cdd-O2d)CtCsCs L8: Cs-(Cds-Cdd-S2d)CtCsCs L8: Cs-(Cds-Cdd-Cd)CtCsCs L5: Cs-CbCdsCsCs L6: Cs-(Cds-O2d)CbCsCs L6: Cs-(Cds-Cd)CbCsCs L7: Cs-(Cds-Cds)CbCsCs L7: Cs-(Cds-Cdd)CbCsCs L8: Cs-(Cds-Cdd-O2d)CbCsCs L8: Cs-(Cds-Cdd-S2d)CbCsCs L8: Cs-(Cds-Cdd-Cd)CbCsCs L5: Cs-CtCtCsCs L6: Cs-(CtN3t)(CtN3t)CsCs L5: Cs-CbCtCsCs L5: Cs-CbCbCsCs L5: Cs-CdsCdsCdsCs L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cs L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Cs L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Cs L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cs L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cs L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Cs L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Cs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cs L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Cs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Cs L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cs L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Cs L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cs L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Cs L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Cs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs L5: Cs-CtCdsCdsCs L6: Cs-(Cds-O2d)(Cds-O2d)CtCs L6: Cs-(Cds-O2d)(Cds-Cd)CtCs L7: Cs-(Cds-O2d)(Cds-Cds)CtCs L7: Cs-(Cds-O2d)(Cds-Cdd)CtCs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CtCs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCs L6: Cs-(Cds-Cd)(Cds-Cd)CtCs L7: Cs-(Cds-Cds)(Cds-Cds)CtCs L7: Cs-(Cds-Cdd)(Cds-Cds)CtCs L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCs L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CtCs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCs L7: Cs-(Cds-Cdd)(Cds-Cdd)CtCs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtCs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtCs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtCs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtCs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCs L5: Cs-CbCdsCdsCs L6: Cs-(Cds-O2d)(Cds-O2d)CbCs L6: Cs-(Cds-O2d)(Cds-Cd)CbCs L7: Cs-(Cds-O2d)(Cds-Cds)CbCs L7: Cs-(Cds-O2d)(Cds-Cdd)CbCs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCs L6: Cs-(Cds-Cd)(Cds-Cd)CbCs L7: Cs-(Cds-Cds)(Cds-Cds)CbCs L7: Cs-(Cds-Cdd)(Cds-Cds)CbCs L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCs L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCs L7: Cs-(Cds-Cdd)(Cds-Cdd)CbCs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCs L5: Cs-CtCtCdsCs L6: Cs-(Cds-O2d)CtCtCs L6: Cs-(Cds-Cd)CtCtCs L7: Cs-(Cds-Cds)CtCtCs L7: Cs-(Cds-Cdd)CtCtCs L8: Cs-(Cds-Cdd-O2d)CtCtCs L8: Cs-(Cds-Cdd-S2d)CtCtCs L8: Cs-(Cds-Cdd-Cd)CtCtCs L5: Cs-CbCtCdsCs L6: Cs-(Cds-O2d)CbCtCs L6: Cs-(Cds-Cd)CbCtCs L7: Cs-(Cds-Cds)CbCtCs L7: Cs-(Cds-Cdd)CbCtCs L8: Cs-(Cds-Cdd-O2d)CbCtCs L8: Cs-(Cds-Cdd-S2d)CbCtCs L8: Cs-(Cds-Cdd-Cd)CbCtCs L5: Cs-CbCbCdsCs L6: Cs-(Cds-O2d)CbCbCs L6: Cs-(Cds-Cd)CbCbCs L7: Cs-(Cds-Cds)CbCbCs L7: Cs-(Cds-Cdd)CbCbCs L8: Cs-(Cds-Cdd-O2d)CbCbCs L8: Cs-(Cds-Cdd-S2d)CbCbCs L8: Cs-(Cds-Cdd-Cd)CbCbCs L5: Cs-CtCtCtCs L5: Cs-CbCtCtCs L5: Cs-CbCbCtCs L5: Cs-CbCbCbCs L5: Cs-CdsCdsCdsCds L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-O2d) L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cd) L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds) L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd) L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d) L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd) L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)(Cds-Cd) L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds) L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)(Cds-Cds) L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds) L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds) L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)(Cds-Cdd) L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)(Cds-Cd) L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds) L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd) L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d) L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd) L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd)(Cds-Cdd) L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd) L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)(Cds-Cd) L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds) L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd) L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d) L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d) L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd) L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)(Cds-Cdd) L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d) L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd) L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd) L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d) L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd) L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd) L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d) L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd) L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d) L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd) L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L5: Cs-CtCdsCdsCds L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Ct L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Ct L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Ct L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Ct L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Ct L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Ct L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Ct L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Ct L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Ct L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Ct L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Ct L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Ct L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Ct L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Ct L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Ct L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Ct L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct L5: Cs-CbCdsCdsCds L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cb L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Cb L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cb L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Cb L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cb L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cb L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Cb L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Cb L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cb L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cb L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Cb L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Cb L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cb L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cb L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Cb L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cb L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Cb L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Cb L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb L5: Cs-CtCtCdsCds L6: Cs-(Cds-O2d)(Cds-O2d)CtCt L6: Cs-(Cds-O2d)(Cds-Cd)CtCt L7: Cs-(Cds-O2d)(Cds-Cds)CtCt L7: Cs-(Cds-O2d)(Cds-Cdd)CtCt L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CtCt L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCt L6: Cs-(Cds-Cd)(Cds-Cd)CtCt L7: Cs-(Cds-Cds)(Cds-Cds)CtCt L7: Cs-(Cds-Cdd)(Cds-Cds)CtCt L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCt L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CtCt L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCt L7: Cs-(Cds-Cdd)(Cds-Cdd)CtCt L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtCt L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtCt L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtCt L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtCt L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCt L5: Cs-CbCtCdsCds L6: Cs-(Cds-O2d)(Cds-O2d)CbCt L6: Cs-(Cds-O2d)(Cds-Cd)CbCt L7: Cs-(Cds-O2d)(Cds-Cds)CbCt L7: Cs-(Cds-O2d)(Cds-Cdd)CbCt L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCt L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCt L6: Cs-(Cds-Cd)(Cds-Cd)CbCt L7: Cs-(Cds-Cds)(Cds-Cds)CbCt L7: Cs-(Cds-Cdd)(Cds-Cds)CbCt L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCt L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCt L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCt L7: Cs-(Cds-Cdd)(Cds-Cdd)CbCt L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCt L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCt L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCt L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCt L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCt L5: Cs-CbCbCdsCds L6: Cs-(Cds-O2d)(Cds-O2d)CbCb L6: Cs-(Cds-O2d)(Cds-Cd)CbCb L7: Cs-(Cds-O2d)(Cds-Cds)CbCb L7: Cs-(Cds-O2d)(Cds-Cdd)CbCb L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCb L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCb L6: Cs-(Cds-Cd)(Cds-Cd)CbCb L7: Cs-(Cds-Cds)(Cds-Cds)CbCb L7: Cs-(Cds-Cdd)(Cds-Cds)CbCb L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCb L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCb L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCb L7: Cs-(Cds-Cdd)(Cds-Cdd)CbCb L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCb L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCb L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCb L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCb L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCb L5: Cs-CtCtCtCds L6: Cs-(Cds-O2d)CtCtCt L6: Cs-(Cds-Cd)CtCtCt L7: Cs-(Cds-Cds)CtCtCt L7: Cs-(Cds-Cdd)CtCtCt L8: Cs-(Cds-Cdd-O2d)CtCtCt L8: Cs-(Cds-Cdd-S2d)CtCtCt L8: Cs-(Cds-Cdd-Cd)CtCtCt L5: Cs-CbCtCtCds L6: Cs-(Cds-O2d)CbCtCt L6: Cs-(Cds-Cd)CbCtCt L7: Cs-(Cds-Cds)CbCtCt L7: Cs-(Cds-Cdd)CbCtCt L8: Cs-(Cds-Cdd-O2d)CbCtCt L8: Cs-(Cds-Cdd-S2d)CbCtCt L8: Cs-(Cds-Cdd-Cd)CbCtCt L5: Cs-CbCbCtCds L6: Cs-(Cds-O2d)CbCbCt L6: Cs-(Cds-Cd)CbCbCt L7: Cs-(Cds-Cds)CbCbCt L7: Cs-(Cds-Cdd)CbCbCt L8: Cs-(Cds-Cdd-O2d)CbCbCt L8: Cs-(Cds-Cdd-S2d)CbCbCt L8: Cs-(Cds-Cdd-Cd)CbCbCt L5: Cs-CbCbCbCds L6: Cs-(Cds-O2d)CbCbCb L6: Cs-(Cds-Cd)CbCbCb L7: Cs-(Cds-Cds)CbCbCb L7: Cs-(Cds-Cdd)CbCbCb L8: Cs-(Cds-Cdd-O2d)CbCbCb L8: Cs-(Cds-Cdd-S2d)CbCbCb L8: Cs-(Cds-Cdd-Cd)CbCbCb L5: Cs-CtCtCtCt L5: Cs-CbCtCtCt L5: Cs-CbCbCtCt L5: Cs-CbCbCbCt L5: Cs-CbCbCbCb L5: Cs-C=SCbCtCt L5: Cs-C=S(Cds-Cd)(Cds-Cd)(Cds-Cd) L6: Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd) L7: Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd) L7: Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d) L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)(Cds-Cdd) L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d) L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd) L6: Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cds) L6: Cs-C=S(Cds-Cds)(Cds-Cdd)(Cds-Cdd) L7: Cs-C=S(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d) L7: Cs-C=S(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd) L7: Cs-C=S(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd) L5: Cs-C=S(Cds-Cd)CtCt L6: Cs-C=S(Cds-Cds)CtCt L6: Cs-C=S(Cds-Cdd)CtCt L7: Cs-C=S(Cds-Cdd-S2d)CtCt L7: Cs-C=S(Cds-Cdd-Cd)CtCt L5: Cs-C=S(Cds-Cd)CtCs L6: Cs-C=S(Cds-Cds)CtCs L6: Cs-C=S(Cds-Cdd)CtCs L7: Cs-C=S(Cds-Cdd-S2d)CtCs L7: Cs-C=S(Cds-Cdd-Cd)CtCs L5: Cs-C=SCbCbCt L5: Cs-C=SCbCsCs L5: Cs-C=SCbCbCs L5: Cs-C=SCtCtCt L5: Cs-C=S(Cds-Cd)(Cds-Cd)Cs L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)Cs L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs L6: Cs-C=S(Cds-Cds)(Cds-Cds)Cs L6: Cs-C=S(Cds-Cdd)(Cds-Cds)Cs L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Cs L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Cs L5: Cs-C=SC=SCtCt L5: Cs-C=SCsCsCs L5: Cs-C=SCtCtCs L5: Cs-C=SC=SC=SCt L5: Cs-C=SC=SC=SCs L5: Cs-C=SC=SC=SC=S L5: Cs-C=SCtCsCs L5: Cs-C=SC=SC=SCb L5: Cs-C=SC=SC=S(Cds-Cd) L6: Cs-C=SC=SC=S(Cds-Cdd) L7: Cs-C=SC=SC=S(Cds-Cdd-Cd) L7: Cs-C=SC=SC=S(Cds-Cdd-S2d) L6: Cs-C=SC=SC=S(Cds-Cds) L5: Cs-C=S(Cds-Cd)(Cds-Cd)Ct L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)Ct L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct L6: Cs-C=S(Cds-Cds)(Cds-Cds)Ct L6: Cs-C=S(Cds-Cdd)(Cds-Cds)Ct L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Ct L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Ct L5: Cs-C=SC=SCtCs L5: Cs-C=SC=SCbCb L5: Cs-C=S(Cds-Cd)CsCs L6: Cs-C=S(Cds-Cds)CsCs L6: Cs-C=S(Cds-Cdd)CsCs L7: Cs-C=S(Cds-Cdd-Cd)CsCs L7: Cs-C=S(Cds-Cdd-S2d)CsCs L5: Cs-C=SC=SCbCt L5: Cs-C=S(Cds-Cd)CbCt L6: Cs-C=S(Cds-Cds)CbCt L6: Cs-C=S(Cds-Cdd)CbCt L7: Cs-C=S(Cds-Cdd-S2d)CbCt L7: Cs-C=S(Cds-Cdd-Cd)CbCt L5: Cs-C=SC=SCsCs L5: Cs-C=S(Cds-Cd)CbCb L6: Cs-C=S(Cds-Cds)CbCb L6: Cs-C=S(Cds-Cdd)CbCb L7: Cs-C=S(Cds-Cdd-S2d)CbCb L7: Cs-C=S(Cds-Cdd-Cd)CbCb L5: Cs-C=SC=S(Cds-Cd)Ct L6: Cs-C=SC=S(Cds-Cds)Ct L6: Cs-C=SC=S(Cds-Cdd)Ct L7: Cs-C=SC=S(Cds-Cdd-Cd)Ct L7: Cs-C=SC=S(Cds-Cdd-S2d)Ct L5: Cs-C=SC=S(Cds-Cd)Cs L6: Cs-C=SC=S(Cds-Cds)Cs L6: Cs-C=SC=S(Cds-Cdd)Cs L7: Cs-C=SC=S(Cds-Cdd-S2d)Cs L7: Cs-C=SC=S(Cds-Cdd-Cd)Cs L5: Cs-C=SC=S(Cds-Cd)(Cds-Cd) L6: Cs-C=SC=S(Cds-Cdd)(Cds-Cds) L7: Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cds) L7: Cs-C=SC=S(Cds-Cdd-Cd)(Cds-Cds) L6: Cs-C=SC=S(Cds-Cdd)(Cds-Cdd) L7: Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cdd-S2d) L7: Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cdd-Cd) L7: Cs-C=SC=S(Cds-Cdd-Cd)(Cds-Cdd-Cd) L6: Cs-C=SC=S(Cds-Cds)(Cds-Cds) L5: Cs-C=SC=S(Cds-Cd)Cb L6: Cs-C=SC=S(Cds-Cdd)Cb L7: Cs-C=SC=S(Cds-Cdd-S2d)Cb L7: Cs-C=SC=S(Cds-Cdd-Cd)Cb L6: Cs-C=SC=S(Cds-Cds)Cb L5: Cs-C=SCbCtCs L5: Cs-C=S(Cds-Cd)CbCs L6: Cs-C=S(Cds-Cds)CbCs L6: Cs-C=S(Cds-Cdd)CbCs L7: Cs-C=S(Cds-Cdd-S2d)CbCs L7: Cs-C=S(Cds-Cdd-Cd)CbCs L5: Cs-C=S(Cds-Cd)(Cds-Cd)Cb L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)Cb L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb L6: Cs-C=S(Cds-Cds)(Cds-Cds)Cb L6: Cs-C=S(Cds-Cdd)(Cds-Cds)Cb L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Cb L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Cb L5: Cs-C=SCbCbCb L5: Cs-C=SC=SCbCs L4: Cs-CCCOs L5: Cs-CsCsCsOs L5: Cs-CdsCsCsOs L6: Cs-(Cds-O2d)CsCsOs L6: Cs-(Cds-Cd)CsCsOs L7: Cs-(Cds-Cds)CsCsOs L7: Cs-(Cds-Cdd)CsCsOs L8: Cs-(Cds-Cdd-O2d)CsCsOs L8: Cs-(Cds-Cdd-Cd)CsCsOs L5: Cs-OsCtCsCs L5: Cs-CbCsCsOs L5: Cs-CdsCdsCsOs L6: Cs-(Cds-O2d)(Cds-O2d)CsOs L6: Cs-(Cds-O2d)(Cds-Cd)CsOs L7: Cs-(Cds-O2d)(Cds-Cds)CsOs L7: Cs-(Cds-O2d)(Cds-Cdd)CsOs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CsOs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CsOs L6: Cs-(Cds-Cd)(Cds-Cd)CsOs L7: Cs-(Cds-Cds)(Cds-Cds)CsOs L7: Cs-(Cds-Cdd)(Cds-Cds)CsOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CsOs L7: Cs-(Cds-Cdd)(Cds-Cdd)CsOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsOs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsOs L5: Cs-CtCdsCsOs L6: Cs-(Cds-O2d)CtCsOs L6: Cs-(Cds-Cd)CtCsOs L7: Cs-(Cds-Cds)CtCsOs L7: Cs-(Cds-Cdd)CtCsOs L8: Cs-(Cds-Cdd-O2d)CtCsOs L8: Cs-(Cds-Cdd-Cd)CtCsOs L5: Cs-CbCdsCsOs L6: Cs-(Cds-O2d)CbCsOs L6: Cs-(Cds-Cd)CbCsOs L7: Cs-(Cds-Cds)CbCsOs L7: Cs-(Cds-Cdd)CbCsOs L8: Cs-(Cds-Cdd-O2d)CbCsOs L8: Cs-(Cds-Cdd-Cd)CbCsOs L5: Cs-CtCtCsOs L5: Cs-CbCtCsOs L5: Cs-CbCbCsOs L5: Cs-CdsCdsCdsOs L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)O2s L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)O2s L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)O2s L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)O2s L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)O2s L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)O2s L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)O2s L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)O2s L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)O2s L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)O2s L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)O2s L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)O2s L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)O2s L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)O2s L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s L5: Cs-CtCdsCdsOs L6: Cs-(Cds-O2d)(Cds-O2d)CtOs L6: Cs-(Cds-O2d)(Cds-Cd)CtOs L7: Cs-(Cds-O2d)(Cds-Cds)CtOs L7: Cs-(Cds-O2d)(Cds-Cdd)CtOs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CtOs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CtOs L6: Cs-(Cds-Cd)(Cds-Cd)CtOs L7: Cs-(Cds-Cds)(Cds-Cds)CtOs L7: Cs-(Cds-Cdd)(Cds-Cds)CtOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CtOs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtOs L7: Cs-(Cds-Cdd)(Cds-Cdd)CtOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtOs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtOs L5: Cs-CbCdsCdsOs L6: Cs-(Cds-O2d)(Cds-O2d)CbOs L6: Cs-(Cds-O2d)(Cds-Cd)CbOs L7: Cs-(Cds-O2d)(Cds-Cds)CbOs L7: Cs-(Cds-O2d)(Cds-Cdd)CbOs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbOs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbOs L6: Cs-(Cds-Cd)(Cds-Cd)CbOs L7: Cs-(Cds-Cds)(Cds-Cds)CbOs L7: Cs-(Cds-Cdd)(Cds-Cds)CbOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbOs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbOs L7: Cs-(Cds-Cdd)(Cds-Cdd)CbOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbOs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbOs L5: Cs-CtCtCdsOs L6: Cs-(Cds-O2d)CtCtOs L6: Cs-(Cds-Cd)CtCtOs L7: Cs-(Cds-Cds)CtCtOs L7: Cs-(Cds-Cdd)CtCtOs L8: Cs-(Cds-Cdd-O2d)CtCtOs L8: Cs-(Cds-Cdd-Cd)CtCtOs L5: Cs-CbCtCdsOs L6: Cs-(Cds-O2d)CbCtOs L6: Cs-(Cds-Cd)CbCtOs L7: Cs-(Cds-Cds)CbCtOs L7: Cs-(Cds-Cdd)CbCtOs L8: Cs-(Cds-Cdd-O2d)CbCtOs L8: Cs-(Cds-Cdd-Cd)CbCtOs L5: Cs-CbCbCdsOs L6: Cs-(Cds-O2d)CbCbOs L6: Cs-(Cds-Cd)CbCbOs L7: Cs-(Cds-Cds)CbCbOs L7: Cs-(Cds-Cdd)CbCbOs L8: Cs-(Cds-Cdd-O2d)CbCbOs L8: Cs-(Cds-Cdd-Cd)CbCbOs L5: Cs-CtCtCtOs L5: Cs-CbCtCtOs L5: Cs-CbCbCtOs L5: Cs-CbCbCbOs L4: Cs-CCOsOs L5: Cs-CsCsOsOs L5: Cs-CdsCsOsOs L6: Cs-(Cds-O2d)CsOsOs L6: Cs-(Cds-Cd)CsOsOs L7: Cs-(Cds-Cds)CsOsOs L7: Cs-(Cds-Cdd)CsOsOs L8: Cs-(Cds-Cdd-O2d)CsOsOs L8: Cs-(Cds-Cdd-Cd)CsOsOs L5: Cs-CdsCdsOsOs L6: Cs-(Cds-O2d)(Cds-O2d)OsOs L6: Cs-(Cds-O2d)(Cds-Cd)OsOs L7: Cs-(Cds-O2d)(Cds-Cds)OsOs L7: Cs-(Cds-O2d)(Cds-Cdd)OsOs L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)OsOs L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)OsOs L6: Cs-(Cds-Cd)(Cds-Cd)OsOs L7: Cs-(Cds-Cds)(Cds-Cds)OsOs L7: Cs-(Cds-Cdd)(Cds-Cds)OsOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)OsOs L7: Cs-(Cds-Cdd)(Cds-Cdd)OsOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)OsOs L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)OsOs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsOs L5: Cs-CtCsOsOs L5: Cs-CtCdsOsOs L6: Cs-(Cds-O2d)CtOsOs L6: Cs-(Cds-Cd)CtOsOs L7: Cs-(Cds-Cds)CtOsOs L7: Cs-(Cds-Cdd)CtOsOs L8: Cs-(Cds-Cdd-O2d)CtOsOs L8: Cs-(Cds-Cdd-Cd)CtOsOs L5: Cs-CtCtOsOs L5: Cs-CbCsOsOs L5: Cs-CbCdsOsOs L6: Cs-(Cds-O2d)CbOsOs L6: Cs-(Cds-Cd)CbOsOs L7: Cs-(Cds-Cds)CbOsOs L7: Cs-(Cds-Cdd)CbOsOs L8: Cs-(Cds-Cdd-O2d)CbOsOs L8: Cs-(Cds-Cdd-Cd)CbOsOs L5: Cs-CbCtOsOs L5: Cs-CbCbOsOs L4: Cs-COsOsOs L5: Cs-CsOsOsOs L5: Cs-CdsOsOsOs L6: Cs-(Cds-O2d)OsOsOs L6: Cs-(Cds-Cd)OsOsOs L7: Cs-(Cds-Cds)OsOsOs L7: Cs-(Cds-Cdd)OsOsOs L8: Cs-(Cds-Cdd-O2d)OsOsOs L8: Cs-(Cds-Cdd-Cd)OsOsOs L5: Cs-CtOsOsOs L5: Cs-CbOsOsOs L4: Cs-OsOsOsOs L4: Cs-COsOsH L5: Cs-CsOsOsH L5: Cs-CdsOsOsH L6: Cs-(Cds-O2d)OsOsH L6: Cs-(Cds-Cd)OsOsH L7: Cs-(Cds-Cds)OsOsH L7: Cs-(Cds-Cdd)OsOsH L8: Cs-(Cds-Cdd-O2d)OsOsH L8: Cs-(Cds-Cdd-Cd)OsOsH L5: Cs-CtOsOsH L5: Cs-CbOsOsH L4: Cs-COsSsH L5: Cs-CsOsSsH L5: Cs-CdsOsSsH L5: Cs-CtOsSsH L5: Cs-CbOsSsH L4: Cs-CCOsSs L5: Cs-CsCsOsSs L4: Cs-COsOsSs L5: Cs-CsOsOsSs L4: Cs-CCOsH L5: Cs-CsCsOsH L5: Cs-CdsCsOsH L6: Cs-(Cds-O2d)CsOsH L6: Cs-(Cds-Cd)CsOsH L7: Cs-(Cds-Cds)CsOsH L7: Cs-(Cds-Cdd)CsOsH L8: Cs-(Cds-Cdd-O2d)CsOsH L8: Cs-(Cds-Cdd-Cd)CsOsH L5: Cs-CdsCdsOsH L6: Cs-(Cds-O2d)(Cds-O2d)OsH L6: Cs-(Cds-O2d)(Cds-Cd)OsH L7: Cs-(Cds-O2d)(Cds-Cds)OsH L7: Cs-(Cds-O2d)(Cds-Cdd)OsH L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)OsH L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)OsH L6: Cs-(Cds-Cd)(Cds-Cd)OsH L7: Cs-(Cds-Cds)(Cds-Cds)OsH L7: Cs-(Cds-Cdd)(Cds-Cds)OsH L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)OsH L7: Cs-(Cds-Cdd)(Cds-Cdd)OsH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)OsH L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)OsH L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsH L5: Cs-CtCsOsH L5: Cs-CtCdsOsH L6: Cs-(Cds-O2d)CtOsH L6: Cs-(Cds-Cd)CtOsH L7: Cs-(Cds-Cds)CtOsH L7: Cs-(Cds-Cdd)CtOsH L8: Cs-(Cds-Cdd-O2d)CtOsH L8: Cs-(Cds-Cdd-Cd)CtOsH L5: Cs-CtCtOsH L5: Cs-CbCsOsH L5: Cs-CbCdsOsH L6: Cs-(Cds-O2d)CbOsH L6: Cs-(Cds-Cd)CbOsH L7: Cs-(Cds-Cds)CbOsH L7: Cs-(Cds-Cdd)CbOsH L8: Cs-(Cds-Cdd-O2d)CbOsH L8: Cs-(Cds-Cdd-Cd)CbOsH L5: Cs-CbCtOsH L5: Cs-CbCbOsH L4: Cs-COsHH L5: Cs-CsOsHH L5: Cs-CdsOsHH L6: Cs-(Cds-O2d)OsHH L6: Cs-(Cds-Cd)OsHH L7: Cs-(Cds-Cds)OsHH L7: Cs-(Cds-Cdd)OsHH L8: Cs-(Cds-Cdd-O2d)OsHH L8: Cs-(Cds-Cdd-Cd)OsHH L5: Cs-CtOsHH L5: Cs-CbOsHH L4: Cs-CCCSs L5: Cs-CsCsCsSs L5: Cs-CdsCsCsSs L6: Cs-(Cds-Cd)CsCsSs L7: Cs-(Cds-Cds)CsCsSs L7: Cs-(Cds-Cdd)CsCsSs L8: Cs-(Cds-Cdd-S2d)CsCsSs L8: Cs-(Cds-Cdd-Cd)CsCsSs L5: Cs-SsCtCsCs L5: Cs-CbCsCsSs L5: Cs-CdsCdsCsSs L6: Cs-(Cds-Cd)(Cds-Cd)CsSs L7: Cs-(Cds-Cds)(Cds-Cds)CsSs L7: Cs-(Cds-Cdd)(Cds-Cds)CsSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CsSs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CsSs L7: Cs-(Cds-Cdd)(Cds-Cdd)CsSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsSs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsSs L5: Cs-CtCdsCsSs L6: Cs-(Cds-Cd)CtCsSs L7: Cs-(Cds-Cds)CtCsSs L7: Cs-(Cds-Cdd)CtCsSs L8: Cs-(Cds-Cdd-S2d)CtCsSs L8: Cs-(Cds-Cdd-Cd)CtCsSs L5: Cs-CbCdsCsSs L6: Cs-(Cds-Cd)CbCsSs L7: Cs-(Cds-Cds)CbCsSs L7: Cs-(Cds-Cdd)CbCsSs L8: Cs-(Cds-Cdd-S2d)CbCsSs L8: Cs-(Cds-Cdd-Cd)CbCsSs L5: Cs-CtCtCsSs L5: Cs-CbCtCsSs L5: Cs-CbCbCsSs L5: Cs-CdsCdsCdsSs L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)S2s L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)S2s L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)S2s L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)S2s L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)S2s L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)S2s L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)S2s L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s L5: Cs-CtCdsCdsSs L6: Cs-(Cds-Cd)(Cds-Cd)CtSs L7: Cs-(Cds-Cds)(Cds-Cds)CtSs L7: Cs-(Cds-Cdd)(Cds-Cds)CtSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CtSs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtSs L7: Cs-(Cds-Cdd)(Cds-Cdd)CtSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtSs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtSs L5: Cs-CbCdsCdsSs L6: Cs-(Cds-Cd)(Cds-Cd)CbSs L7: Cs-(Cds-Cds)(Cds-Cds)CbSs L7: Cs-(Cds-Cdd)(Cds-Cds)CbSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbSs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbSs L7: Cs-(Cds-Cdd)(Cds-Cdd)CbSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbSs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbSs L5: Cs-CtCtCdsSs L6: Cs-(Cds-Cd)CtCtSs L7: Cs-(Cds-Cds)CtCtSs L7: Cs-(Cds-Cdd)CtCtSs L8: Cs-(Cds-Cdd-S2d)CtCtSs L8: Cs-(Cds-Cdd-Cd)CtCtSs L5: Cs-CbCtCdsSs L6: Cs-(Cds-Cd)CbCtSs L7: Cs-(Cds-Cds)CbCtSs L7: Cs-(Cds-Cdd)CbCtSs L8: Cs-(Cds-Cdd-S2d)CbCtSs L8: Cs-(Cds-Cdd-Cd)CbCtSs L5: Cs-CbCbCdsSs L6: Cs-(Cds-Cd)CbCbSs L7: Cs-(Cds-Cds)CbCbSs L7: Cs-(Cds-Cdd)CbCbSs L8: Cs-(Cds-Cdd-S2d)CbCbSs L8: Cs-(Cds-Cdd-Cd)CbCbSs L5: Cs-CtCtCtSs L5: Cs-CbCtCtSs L5: Cs-CbCbCtSs L5: Cs-CbCbCbSs L5: Cs-C=SCbCsSs L5: Cs-C=SCsCsSs L5: Cs-C=S(Cds-Cd)(Cds-Cd)S2s L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)S2s L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s L6: Cs-C=S(Cds-Cdd)(Cds-Cds)S2s L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)S2s L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)S2s L6: Cs-C=S(Cds-Cds)(Cds-Cds)S2s L5: Cs-C=S(Cds-Cd)CtSs L6: Cs-C=S(Cds-Cds)CtSs L6: Cs-C=S(Cds-Cdd)CtSs L7: Cs-C=S(Cds-Cdd-S2d)CtSs L7: Cs-C=S(Cds-Cdd-Cd)CtSs L5: Cs-C=SCtCsSs L5: Cs-C=SC=SC=SSs L5: Cs-C=SC=S(Cds-Cd)S2s L6: Cs-C=SC=S(Cds-Cds)S2s L6: Cs-C=SC=S(Cds-Cdd)S2s L7: Cs-C=SC=S(Cds-Cdd-S2d)S2s L7: Cs-C=SC=S(Cds-Cdd-Cd)S2s L5: Cs-C=SCbCbSs L5: Cs-C=SC=SCbSs L5: Cs-C=SC=SCsSs L5: Cs-C=SCtCtSs L5: Cs-C=S(Cds-Cd)CbSs L6: Cs-C=S(Cds-Cdd)CbSs L7: Cs-C=S(Cds-Cdd-Cd)CbSs L7: Cs-C=S(Cds-Cdd-S2d)CbSs L6: Cs-C=S(Cds-Cds)CbSs L5: Cs-C=SCbCtSs L5: Cs-C=SC=SCtSs L5: Cs-C=S(Cds-Cd)CsSs L6: Cs-C=S(Cds-Cds)CsSs L6: Cs-C=S(Cds-Cdd)CsSs L7: Cs-C=S(Cds-Cdd-S2d)CsSs L7: Cs-C=S(Cds-Cdd-Cd)CsSs L4: Cs-CCSsSs L5: Cs-CsCsSsSs L5: Cs-CdsCsSsSs L6: Cs-(Cds-Cd)CsSsSs L7: Cs-(Cds-Cds)CsSsSs L7: Cs-(Cds-Cdd)CsSsSs L8: Cs-(Cds-Cdd-S2d)CsSsSs L8: Cs-(Cds-Cdd-Cd)CsSsSs L5: Cs-CdsCdsSsSs L6: Cs-(Cds-Cd)(Cds-Cd)SsSs L7: Cs-(Cds-Cds)(Cds-Cds)SsSs L7: Cs-(Cds-Cdd)(Cds-Cds)SsSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)SsSs L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)SsSs L7: Cs-(Cds-Cdd)(Cds-Cdd)SsSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)SsSs L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)SsSs L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)SsSs L5: Cs-CtCsSsSs L5: Cs-CtCdsSsSs L6: Cs-(Cds-Cd)CtSsSs L7: Cs-(Cds-Cds)CtSsSs L7: Cs-(Cds-Cdd)CtSsSs L8: Cs-(Cds-Cdd-S2d)CtSsSs L8: Cs-(Cds-Cdd-Cd)CtSsSs L5: Cs-CtCtSsSs L5: Cs-CbCsSsSs L5: Cs-CbCdsSsSs L6: Cs-(Cds-Cd)CbSsSs L7: Cs-(Cds-Cds)CbSsSs L7: Cs-(Cds-Cdd)CbSsSs L8: Cs-(Cds-Cdd-S2d)CbSsSs L8: Cs-(Cds-Cdd-Cd)CbSsSs L5: Cs-CbCtSsSs L5: Cs-CbCbSsSs L5: Cs-C=SCsSsSs L5: Cs-C=S(Cds-Cd)SsSs L6: Cs-C=S(Cds-Cdd)SsSs L7: Cs-C=S(Cds-Cdd-Cd)SsSs L7: Cs-C=S(Cds-Cdd-S2d)SsSs L6: Cs-C=S(Cds-Cds)SsSs L5: Cs-C=SC=SSsSs L5: Cs-C=SCbSsSs L5: Cs-C=SCtSsSs L4: Cs-CSsSsSs L5: Cs-CsSsSsSs L5: Cs-CdsSsSsSs L6: Cs-(Cds-Cd)SsSsSs L7: Cs-(Cds-Cds)SsSsSs L7: Cs-(Cds-Cdd)SsSsSs L8: Cs-(Cds-Cdd-S2d)SsSsSs L8: Cs-(Cds-Cdd-Cd)SsSsSs L5: Cs-CtSsSsSs L5: Cs-CbSsSsSs L5: Cs-C=SSsSsSs L4: Cs-SsSsSsSs L4: Cs-CSsSsH L5: Cs-CsSsSsH L5: Cs-CdsSsSsH L6: Cs-(Cds-Cd)SsSsH L7: Cs-(Cds-Cds)SsSsH L7: Cs-(Cds-Cdd)SsSsH L8: Cs-(Cds-Cdd-S2d)SsSsH L8: Cs-(Cds-Cdd-Cd)SsSsH L5: Cs-CtSsSsH L5: Cs-CbSsSsH L5: Cs-C=SSsSsH L4: Cs-CCSsH L5: Cs-CsCsSsH L5: Cs-CdsCsSsH L6: Cs-(Cds-Cd)CsSsH L7: Cs-(Cds-Cds)CsSsH L7: Cs-(Cds-Cdd)CsSsH L8: Cs-(Cds-Cdd-S2d)CsSsH L8: Cs-(Cds-Cdd-Cd)CsSsH L5: Cs-CdsCdsSsH L6: Cs-(Cds-Cd)(Cds-Cd)SsH L7: Cs-(Cds-Cds)(Cds-Cds)SsH L7: Cs-(Cds-Cdd)(Cds-Cds)SsH L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)SsH L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)SsH L7: Cs-(Cds-Cdd)(Cds-Cdd)SsH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)SsH L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)SsH L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)SsH L5: Cs-CtCsSsH L5: Cs-CtCdsSsH L6: Cs-(Cds-Cd)CtSsH L7: Cs-(Cds-Cds)CtSsH L7: Cs-(Cds-Cdd)CtSsH L8: Cs-(Cds-Cdd-S2d)CtSsH L8: Cs-(Cds-Cdd-Cd)CtSsH L5: Cs-CtCtSsH L5: Cs-CbCsSsH L5: Cs-CbCdsSsH L6: Cs-(Cds-Cd)CbSsH L7: Cs-(Cds-Cds)CbSsH L7: Cs-(Cds-Cdd)CbSsH L8: Cs-(Cds-Cdd-S2d)CbSsH L8: Cs-(Cds-Cdd-Cd)CbSsH L5: Cs-CbCtSsH L5: Cs-CbCbSsH L5: Cs-C=SCbSsH L5: Cs-C=SC=SSsH L5: Cs-C=SCsSsH L5: Cs-C=SCtSsH L5: Cs-C=S(Cds-Cd)SsH L6: Cs-C=S(Cds-Cdd)SsH L7: Cs-C=S(Cds-Cdd-Cd)SsH L7: Cs-C=S(Cds-Cdd-S2d)SsH L6: Cs-C=S(Cds-Cds)SsH L4: Cs-CSsHH L5: Cs-CsSsHH L5: Cs-CdsSsHH L6: Cs-(Cds-Cd)SsHH L7: Cs-(Cds-Cds)SsHH L7: Cs-(Cds-Cdd)SsHH L8: Cs-(Cds-Cdd-S2d)SsHH L8: Cs-(Cds-Cdd-Cd)SsHH L5: Cs-CtSsHH L5: Cs-CbSsHH L5: Cs-C=SSsHH L2: O L3: O2d L4: O2d-Cd L4: O2d-O2d L4: O2d-N3d L4: O2d-N5dc L3: O2s L4: O2s-N L5: O2s-CN L6: O2s-CsN3s L6: O2s-CsN3d L7: O2s-Cs(N3dOd) L6: O2s-CdN3d L7: O2s-(Cd-Cd)(N3dOd) L6: O2s-CsN5d L7: O2s-Cs(N5dOdOs) L6: O2s-CdN5d L7: O2s-(Cd-CdHH)(N5dOdOs) L5: O2s-ON L6: O2s-OsN3s L6: O2s-OsN3d L7: O2s-O2s(N3dOd) L5: O2s-NN L6: O2s-N3sN3s L6: O2s-N3sN3d L7: O2s-N3s(N3dOd) L4: O2s-HH L4: O2s-OsH L4: O2s-OsOs L4: O2s-CH L5: O2s-CtH L5: O2s-CdsH L6: O2s-(Cds-O2d)H L6: O2s-(Cds-Cd)H L5: O2s-CsH L5: O2s-CbH L5: O2s-CSH L4: O2s-OsC L5: O2s-OsCt L5: O2s-OsCds L6: O2s-O2s(Cds-O2d) L6: O2s-O2s(Cds-Cd) L5: O2s-OsCs L5: O2s-OsCb L4: O2s-CC L5: O2s-CtCt L5: O2s-CtCds L6: O2s-Ct(Cds-O2d) L6: O2s-Ct(Cds-Cd) L5: O2s-CtCs L6: O2s-Cs(CtN3t) L5: O2s-CtCb L5: O2s-CdsCds L6: O2s-(Cds-O2d)(Cds-O2d) L6: O2s-(Cds-O2d)(Cds-Cd) L6: O2s-(Cds-Cd)(Cds-Cd) L5: O2s-CdsCs L6: O2s-Cs(Cds-O2d) L6: O2s-Cs(Cds-Cd) L5: O2s-CdsCb L6: O2s-Cb(Cds-O2d) L6: O2s-Cb(Cds-Cd) L5: O2s-CsCs L5: O2s-CsCb L5: O2s-CbCb L5: O2s-Cs(Cds-S2d) L2: Si L2: S L3: S2d L4: S2d-Cd L4: S2d-S2d L3: S2s L4: S2s-HH L4: S2s-CH L5: S2s-CsH L5: S2s-CdH L5: S2s-CtH L5: S2s-CbH L5: S2s-COH L5: S2s-C=SH L4: S2s-SsH L4: S2s-SsSs L4: S2s-SsC L5: S2s-SsCs L5: S2s-SsCd L5: S2s-SsCt L5: S2s-SsCb L5: S2s-C=SSs L4: S2s-CC L5: S2s-CsCs L5: S2s-CsCd L5: S2s-CsCO L5: S2s-CsCt L5: S2s-CsCb L5: S2s-CdCd L5: S2s-CdCt L5: S2s-CdCb L5: S2s-CtCt L5: S2s-CtCb L5: S2s-CbCb L5: S2s-C=SCs L5: S2s-C=SCt L5: S2s-C=SC=S L5: S2s-C=SCd L5: S2s-C=SCb L2: N L3: N1dc L3: N3s L4: N3s-CHH L5: N3s-CsHH L5: N3s-CbHH L5: N3s-(CO)HH L5: N3s-CdHH L4: N3s-CCH L5: N3s-CsCsH L5: N3s-CbCsH L5: N3s-CbCbH L5: N3s-(CO)CsH L5: N3s-(CO)CbH L5: N3s-(CO)(CO)H L5: N3s-(CtN3t)CsH L5: N3s-(CdCd)CsH L4: N3s-CCC L5: N3s-CsCsCs L5: N3s-CbCsCs L5: N3s-(CO)CsCs L5: N3s-(CO)(CO)Cs L5: N3s-(CO)(CO)Cb L5: N3s-(CtN3t)CsCs L5: N3s-(CdCd)CsCs L4: N3s-N3sHH L4: N3s-NCH L5: N3s-N3sCsH L5: N3s-N3sCbH L5: N3s-CsH(N3dOd) L5: N3s-CsH(N5dOdOs) L5: N3s-(CdCd)HN3s L4: N3s-NCC L5: N3s-NCsCs L6: N3s-CsCsN3s L6: N3s-CsCs(N3dOd) L6: N3s-CsCs(N5dOdOs) L5: N3s-NCdCs L6: N3s-(CdCd)CsN3s L4: N3s-CsHOs L4: N3s-CsCsOs L4: N3s-OsHH L3: N3d L4: N3d-CdH L4: N3d-N3dH L4: N3d-N3dN3s L4: N3d-OdOs L4: N3d-OdN3s L4: N3d-CsR L5: N3d-OdC L5: N3d-CdCs L5: N3d-N3dCs L4: N3d-CbR L3: N5dc L4: N5dc-OdOsCs L4: N5dc-OdOsCd L4: N5dc-OdOsOs L4: N5dc-OdOsN3s L3: N5ddc """ )
18.390346
120
0.431145
161,696
853,073
2.274626
0.013847
0.058932
0.072116
0.033459
0.905622
0.882612
0.857571
0.837655
0.807389
0.79256
0
0.165521
0.319701
853,073
46,386
121
18.390743
0.468237
0.000042
0
0.611813
0
0.005942
0.21866
0.106877
0
0
0
0
0
1
0
false
0
0
0
0
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
d175957c0538886602a21e033a2262d8ddcf487d
429
py
Python
Pi to the nth/Pi.py
huffman442/Projects
16dfcd9bff6e283d6113ed6ebda3f8127576457a
[ "MIT" ]
null
null
null
Pi to the nth/Pi.py
huffman442/Projects
16dfcd9bff6e283d6113ed6ebda3f8127576457a
[ "MIT" ]
null
null
null
Pi to the nth/Pi.py
huffman442/Projects
16dfcd9bff6e283d6113ed6ebda3f8127576457a
[ "MIT" ]
null
null
null
import sys print("How many decimal places would you like to calculate pi to(up to 250 places)?") digits = input() + 2 pi = "3.1415926535897932384626433832795028841971693993751058209749445923078164062862089986280348253421170679821480865132823066470938446095505822317253594081284811174502841027019385211055596446229489549303819644288109756659334461284756482337867831652712019091" for x in range(digits): sys.stdout.write(pi[x])
71.5
259
0.869464
34
429
10.970588
0.764706
0
0
0
0
0
0
0
0
0
0
0.642317
0.074592
429
6
260
71.5
0.297229
0
0
0
0
0
0.762791
0.586047
0
1
0
0
0
1
0
false
0
0.166667
0
0.166667
0.166667
0
0
1
null
0
0
0
0
0
0
0
0
0
0
1
0
0
1
0
0
1
0
0
0
0
0
1
1
null
1
0
0
0
0
0
0
0
0
0
0
0
0
7
0f1d36f35254ddeffdef80a6fff0825f7b6ac405
352
py
Python
Pikachurun.py
kartikay22/Pikachu-X
a8bc32192bbfdf0af5dc9ced1194901e4301c7da
[ "MIT" ]
null
null
null
Pikachurun.py
kartikay22/Pikachu-X
a8bc32192bbfdf0af5dc9ced1194901e4301c7da
[ "MIT" ]
null
null
null
Pikachurun.py
kartikay22/Pikachu-X
a8bc32192bbfdf0af5dc9ced1194901e4301c7da
[ "MIT" ]
null
null
null
##os.system("git clone https://github.com/kartikay22/Pikachu-X /root/userbot && mkdir /root/userbot/bin/ && cd /root/userbot/ && chmod +x /usr/local/bin/* && python3 -m userbot") import os os.system("git clone https://github.com/kartikay22/Pikachu-X && mkdir /root/userbot/bin/ && cd /root/userbot/ && chmod +x /usr/local/bin/* && python3 -m userbot")
88
178
0.690341
54
352
4.5
0.388889
0.226337
0.090535
0.131687
0.921811
0.921811
0.921811
0.921811
0.921811
0.921811
0
0.019048
0.105114
352
3
179
117.333333
0.752381
0.5
0
0
0
0.5
0.856322
0
0
0
0
0
0
1
0
true
0
0.5
0
0.5
0
0
0
0
null
1
0
0
1
1
1
1
1
1
0
0
0
0
1
0
1
0
0
0
0
0
1
1
0
null
0
0
0
0
0
0
1
0
1
0
0
0
0
13
0f57cb2f34e331427865a21c3c25833c3498f6cf
87
py
Python
Problems/Working with strings/task.py
gabrielizalo/jetbrains-academy-python-credit-calculator
093bffde653510348adb35188ecf5cb99e11e73e
[ "MIT" ]
null
null
null
Problems/Working with strings/task.py
gabrielizalo/jetbrains-academy-python-credit-calculator
093bffde653510348adb35188ecf5cb99e11e73e
[ "MIT" ]
null
null
null
Problems/Working with strings/task.py
gabrielizalo/jetbrains-academy-python-credit-calculator
093bffde653510348adb35188ecf5cb99e11e73e
[ "MIT" ]
null
null
null
# put your code here import string print(string.digits) print(string.ascii_lowercase)
14.5
29
0.804598
13
87
5.307692
0.769231
0.318841
0
0
0
0
0
0
0
0
0
0
0.114943
87
5
30
17.4
0.896104
0.206897
0
0
0
0
0
0
0
0
0
0.2
0
1
0
true
0
0.333333
0
0.333333
0.666667
1
0
0
null
1
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
1
0
0
0
1
0
1
0
0
1
0
7
7e472525076f3ec3aae3c5ce6347ec33c5dfd7de
190,630
py
Python
backend/tracim_backend/tests/functional/test_contents.py
lezardrouge/tracim
713ff6066767554333e7e0b1de608ec1a7e4229c
[ "MIT" ]
null
null
null
backend/tracim_backend/tests/functional/test_contents.py
lezardrouge/tracim
713ff6066767554333e7e0b1de608ec1a7e4229c
[ "MIT" ]
null
null
null
backend/tracim_backend/tests/functional/test_contents.py
lezardrouge/tracim
713ff6066767554333e7e0b1de608ec1a7e4229c
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- import io from urllib.parse import quote from PIL import Image from depot.io.utils import FileIntent import pytest import transaction from tracim_backend.app_models.contents import content_type_list from tracim_backend.error import ErrorCode from tracim_backend.fixtures.content import Content as ContentFixtures from tracim_backend.fixtures.users_and_groups import Base as BaseFixture from tracim_backend.lib.core.content import ContentApi from tracim_backend.lib.core.workspace import WorkspaceApi from tracim_backend.models.auth import User from tracim_backend.models.revision_protection import new_revision from tracim_backend.models.setup_models import get_tm_session from tracim_backend.tests import FunctionalTest from tracim_backend.tests import create_1000px_png_test_image from tracim_backend.tests import set_html_document_slug_to_legacy class TestFolder(FunctionalTest): """ Tests for /api/v2/workspaces/{workspace_id}/folders/{content_id} endpoint """ fixtures = [BaseFixture] def _setup_basics(self) -> None: self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.dbsession = get_tm_session(self.session_factory, transaction.manager) self.admin = self.dbsession.query(User).filter(User.email == "admin@admin.admin").one() self.workspace_api = WorkspaceApi( current_user=self.admin, session=self.dbsession, config=self.app_config ) self.content_api = ContentApi( current_user=self.admin, session=self.dbsession, config=self.app_config ) self.workspace = self.workspace_api.create_workspace(label="test", save_now=True) self.folder = self.content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=self.workspace, do_save=True, do_notify=False, ) transaction.commit() def test_api__get_folder__ok_200__nominal_case(self) -> None: """ Get one folder content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test-folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "test-folder" assert content["parent_id"] is None assert content["show_in_ui"] is True assert content["slug"] == "test-folder" assert content["status"] == "open" assert content["workspace_id"] == test_workspace.workspace_id assert content["current_revision_id"] == folder.revision_id # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"]["user_id"] == 1 assert content["last_modifier"]["public_name"] == "Global manager" assert content["last_modifier"]["avatar_url"] is None assert content["raw_content"] == "" def test_api__get_folder__err_400__wrong_content_type(self) -> None: """ Get one folder of a content content 7 is not folder """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) thread = content_api.create( label="thread", content_type_slug=content_type_list.Thread.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=thread.content_id ), status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_TYPE_NOT_ALLOWED def test_api__get_folder__err_400__content_does_not_exist(self) -> None: """ Get one folder content (content 170 does not exist in db) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) transaction.commit() res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/folders/170".format( workspace_id=test_workspace.workspace_id ), status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND def test_api__get_folder__err_400__content_not_in_workspace(self) -> None: """ Get one folders of a content (content is in another workspace) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) test_workspace2 = workspace_api.create_workspace(label="test2", save_now=True) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace2.workspace_id, content_id=folder.content_id ), status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND def test_api__get_folder__err_400__workspace_does_not_exist(self) -> None: """ Get one folder content (Workspace 40 does not exist) """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/40/folders/{content_id}".format(content_id=folder.content_id), status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.WORKSPACE_NOT_FOUND def test_api__get_folder__err_400__workspace_id_is_not_int(self) -> None: """ Get one folder content, workspace id is not int """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/coucou/folders/{content_id}".format(content_id=folder.content_id), status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.WORKSPACE_INVALID_ID def test_api__get_folder__err_400__content_id_is_not_int(self) -> None: """ Get one folder content, content_id is not int """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/folders/coucou".format( workspace_id=test_workspace.workspace_id ), status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_INVALID_ID def test_api__update_folder__err_400__empty_label(self) -> None: """ Update(put) one folder content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = { "label": "", "raw_content": "<p> Le nouveau contenu </p>", "sub_content_types": [content_type_list.Folder.slug], } res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=400, ) # INFO - G.M - 2018-09-10 - Handled by marshmallow schema assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR def test_api__update_folder__ok_200__nominal_case(self) -> None: """ Update(put) one html document of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = { "label": "My New label", "raw_content": "<p> Le nouveau contenu </p>", "sub_content_types": [content_type_list.Folder.slug], } res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] is None assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == test_workspace.workspace_id assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["sub_content_types"] == [content_type_list.Folder.slug] def test_api__update_folder__err_400__not_modified(self) -> None: """ Update(put) one html document of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = { "label": "My New label", "raw_content": "<p> Le nouveau contenu </p>", "sub_content_types": [content_type_list.Folder.slug], } res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] is None assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == test_workspace.workspace_id assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["sub_content_types"] == [content_type_list.Folder.slug] res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.SAME_VALUE_ERROR def test_api__update_folder__err_400__allowed_content_changed_only(self) -> None: """ Update(put) one folder but change only allowed content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = { "label": "My New label", "raw_content": "<p> Le nouveau contenu </p>", "sub_content_types": [content_type_list.Folder.slug], } res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] is None assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == test_workspace.workspace_id assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["sub_content_types"] == [content_type_list.Folder.slug] params = { "label": "My New label", "raw_content": "<p> Le nouveau contenu </p>", "sub_content_types": [content_type_list.Folder.slug, content_type_list.Thread.slug], } res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] is None assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == test_workspace.workspace_id assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert set(content["sub_content_types"]) == set( [content_type_list.Folder.slug, content_type_list.Thread.slug] ) def test_api__update_folder__err_400__label_changed_only(self) -> None: """ Update(put) one folder but change only allowed content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = { "label": "My New label", "raw_content": "<p> Le nouveau contenu </p>", "sub_content_types": [content_type_list.Folder.slug], } res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] is None assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == test_workspace.workspace_id assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["sub_content_types"] == [content_type_list.Folder.slug] params = { "label": "My New label 2", "raw_content": "<p> Le nouveau contenu </p>", "sub_content_types": [content_type_list.Folder.slug], } res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label 2" assert content["parent_id"] is None assert content["show_in_ui"] is True assert content["slug"] == "my-new-label-2" assert content["status"] == "open" assert content["workspace_id"] == test_workspace.workspace_id assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert set(content["sub_content_types"]) == set([content_type_list.Folder.slug]) def test_api__update_folder__err_400__label_already_used(self) -> None: """ Update(put) one html document of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) content_api.create( label="already_used", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = { "label": "already_used", "raw_content": "<p> Le nouveau contenu </p>", "sub_content_types": [content_type_list.Folder.slug], } res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=400, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER def test_api__get_folder_revisions__ok_200__nominal_case(self) -> None: """ Get one html document of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test-folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=folder): content_api.update_content( folder, new_label="test-folder-updated", new_content="Just a test" ) content_api.save(folder) with new_revision(session=dbsession, tm=transaction.manager, content=folder): content_api.archive(folder) content_api.save(folder) with new_revision(session=dbsession, tm=transaction.manager, content=folder): content_api.unarchive(folder) content_api.save(folder) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/folders/{content_id}/revisions".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), status=200, ) revisions = res.json_body assert len(revisions) == 4 revision = revisions[0] assert revision["content_type"] == "folder" assert revision["content_id"] == folder.content_id assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is False assert revision["label"] == "test-folder" assert revision["parent_id"] is None assert revision["show_in_ui"] is True assert revision["slug"] == "test-folder" assert revision["status"] == "open" assert revision["workspace_id"] == test_workspace.workspace_id assert revision["revision_id"] assert revision["revision_type"] == "creation" assert revision["sub_content_types"] # TODO - G.M - 2018-06-173 - Test with real comments assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 1 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Global manager" revision = revisions[1] assert revision["content_type"] == "folder" assert revision["content_id"] == folder.content_id assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is False assert revision["label"] == "test-folder-updated" assert revision["parent_id"] is None assert revision["show_in_ui"] is True assert revision["slug"] == "test-folder-updated" assert revision["status"] == "open" assert revision["workspace_id"] == test_workspace.workspace_id assert revision["revision_id"] assert revision["revision_type"] == "edition" assert revision["sub_content_types"] # TODO - G.M - 2018-06-173 - Test with real comments assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 1 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Global manager" revision = revisions[2] assert revision["content_type"] == "folder" assert revision["content_id"] == folder.content_id assert revision["is_archived"] is True assert revision["is_deleted"] is False assert revision["is_editable"] is False assert revision["label"] != "test-folder-updated" assert revision["label"].startswith("test-folder-updated") assert revision["parent_id"] is None assert revision["show_in_ui"] is True assert revision["slug"] != "test-folder-updated" assert revision["slug"].startswith("test-folder-updated") assert revision["status"] == "open" assert revision["workspace_id"] == test_workspace.workspace_id assert revision["revision_id"] assert revision["revision_type"] == "archiving" assert revision["sub_content_types"] # TODO - G.M - 2018-06-173 - Test with real comments assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 1 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Global manager" revision = revisions[3] assert revision["content_type"] == "folder" assert revision["content_id"] == folder.content_id assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is True assert revision["label"].startswith("test-folder-updated") assert revision["parent_id"] is None assert revision["show_in_ui"] is True assert revision["slug"].startswith("test-folder-updated") assert revision["status"] == "open" assert revision["workspace_id"] == test_workspace.workspace_id assert revision["revision_id"] assert revision["revision_type"] == "unarchiving" assert revision["sub_content_types"] # TODO - G.M - 2018-06-173 - Test with real comments assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 1 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Global manager" def test_api__set_folder_status__ok_200__nominal_case(self) -> None: """ Get one folder content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "closed-deprecated"} # before res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["status"] == "open" # set status self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}/status".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=204, ) # after res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/folders/{content_id}".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), status=200, ) content = res.json_body assert content["content_type"] == "folder" assert content["content_id"] == folder.content_id assert content["status"] == "closed-deprecated" def test_api__set_folder_status__err_400__wrong_status(self) -> None: """ Get one folder content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "unexistant-status"} dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) test_workspace = workspace_api.create_workspace(label="test", save_now=True) folder = content_api.create( label="test_folder", content_type_slug=content_type_list.Folder.slug, workspace=test_workspace, do_save=True, do_notify=False, ) transaction.commit() res = self.testapp.put_json( "/api/v2/workspaces/{workspace_id}/folders/{content_id}/status".format( workspace_id=test_workspace.workspace_id, content_id=folder.content_id ), params=params, status=400, ) # TODO - G.M - 2018-09-10 - handle by marshmallow schema assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR class TestHtmlDocuments(FunctionalTest): """ Tests for /api/v2/workspaces/{workspace_id}/html-documents/{content_id} endpoint """ fixtures = [BaseFixture, ContentFixtures] def test_api__get_html_document__ok_200__legacy_slug(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) set_html_document_slug_to_legacy(self.session_factory) res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "Tiramisu Recipe" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "tiramisu-recipe" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 27 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] != content["author"] assert content["last_modifier"]["user_id"] == 3 assert content["last_modifier"]["public_name"] == "Bob i." assert content["last_modifier"]["avatar_url"] is None assert ( content["raw_content"] == "<p>To cook a great Tiramisu, you need many ingredients.</p>" ) assert content["file_extension"] == ".document.html" def test_api__get_html_document__ok_200__nominal_case(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "Tiramisu Recipe" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "tiramisu-recipe" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 27 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] != content["author"] assert content["last_modifier"]["user_id"] == 3 assert content["last_modifier"]["public_name"] == "Bob i." assert content["last_modifier"]["avatar_url"] is None assert ( content["raw_content"] == "<p>To cook a great Tiramisu, you need many ingredients.</p>" ) assert content["file_extension"] == ".document.html" def test_api__get_html_document__ok_200__archived_content(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put_json("/api/v2/workspaces/2/contents/6/archived", status=204) res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["is_archived"] is True def test_api__get_html_document__ok_200__deleted_content(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put_json("/api/v2/workspaces/2/contents/6/trashed", status=204) res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["is_deleted"] is True def test_api__get_html_document__err_400__wrong_content_type(self) -> None: """ Get one html document of a content content 7 is not html_document """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/html-documents/7", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_TYPE_NOT_ALLOWED def test_api__get_html_document__err_400__content_does_not_exist(self) -> None: """ Get one html document of a content (content 170 does not exist in db """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/html-documents/170", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND def test_api__get_html_document__err_400__content_not_in_workspace(self) -> None: """ Get one html document of a content (content 6 is in workspace 2) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/1/html-documents/6", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND def test_api__get_html_document__err_400__workspace_does_not_exist(self) -> None: """ Get one html document of a content (Workspace 40 does not exist) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/40/html-documents/6", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.WORKSPACE_NOT_FOUND def test_api__get_html_document__err_400__workspace_id_is_not_int(self) -> None: """ Get one html document of a content, workspace id is not int """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/coucou/html-documents/6", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.WORKSPACE_INVALID_ID def test_api__get_html_document__err_400__content_id_is_not_int(self) -> None: """ Get one html document of a content, content_id is not int """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/html-documents/coucou", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_INVALID_ID def test_api__update_html_document__err_400__empty_label(self) -> None: """ Update(put) one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6", params=params, status=400 ) # INFO - G.M - 2018-09-10 - Handled by marshmallow schema assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR def test_api__update_html_document__ok_200__nominal_case(self) -> None: """ Update(put) one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6", params=params, status=200 ) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 28 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["file_extension"] == ".document.html" res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 28 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["file_extension"] == ".document.html" def test_api__update_html_document__err_400__not_editable(self) -> None: """ Update(put) one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "closed-deprecated"} self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6/status", params=params, status=204 ) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu ! </p>"} res = self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6", params=params, status=400 ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE def test_api__update_html_document__err_400__not_modified(self) -> None: """ Update(put) one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6", params=params, status=200 ) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 28 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 28 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" res = self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6", params=params, status=400 ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.SAME_VALUE_ERROR def test_api__get_html_document_revisions__ok_200__nominal_case(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/html-documents/6/revisions", status=200) revisions = res.json_body assert len(revisions) == 3 revision = revisions[0] assert revision["content_type"] == "html-document" assert revision["content_id"] == 6 assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is False assert revision["label"] == "Tiramisu Recipes!!!" assert revision["parent_id"] == 3 assert revision["show_in_ui"] is True assert revision["slug"] == "tiramisu-recipes" assert revision["status"] == "open" assert revision["workspace_id"] == 2 assert revision["revision_id"] == 6 assert revision["revision_type"] == "creation" assert revision["sub_content_types"] # TODO - G.M - 2018-06-173 - Test with real comments assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 1 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Global manager" revision = revisions[1] assert revision["content_type"] == "html-document" assert revision["content_id"] == 6 assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is False assert revision["label"] == "Tiramisu Recipes!!!" assert revision["parent_id"] == 3 assert revision["show_in_ui"] is True assert revision["slug"] == "tiramisu-recipes" assert revision["status"] == "open" assert revision["workspace_id"] == 2 assert revision["revision_id"] == 7 assert revision["revision_type"] == "edition" assert revision["sub_content_types"] # TODO - G.M - 2018-06-173 - Test with real comments assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 1 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Global manager" revision = revisions[2] assert revision["content_type"] == "html-document" assert revision["content_id"] == 6 assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is True assert revision["label"] == "Tiramisu Recipe" assert revision["parent_id"] == 3 assert revision["show_in_ui"] is True assert revision["slug"] == "tiramisu-recipe" assert revision["status"] == "open" assert revision["workspace_id"] == 2 assert revision["revision_id"] == 27 assert revision["revision_type"] == "edition" assert revision["sub_content_types"] # TODO - G.M - 2018-06-173 - Test with real comments assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 3 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Bob i." assert revision["file_extension"] == ".document.html" def test_api__set_html_document_status__ok_200__nominal_case(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "closed-deprecated"} # before res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["status"] == "open" # set status self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6/status", params=params, status=204 ) # after res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200) content = res.json_body assert content["content_type"] == "html-document" assert content["content_id"] == 6 assert content["status"] == "closed-deprecated" def test_api__set_html_document_status__err_400__wrong_status(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "unexistant-status"} res = self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6/status", params=params, status=400 ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR def test_api__set_document_status__err_400__same_status(self) -> None: self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "open"} res = self.testapp.put_json( "/api/v2/workspaces/2/html-documents/6/status", params=params, status=400 ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.INVALID_STATUS_CHANGE class TestFiles(FunctionalTest): """ Tests for /api/v2/workspaces/{workspace_id}/files/{content_id} endpoint """ fixtures = [BaseFixture, ContentFixtures] def test_api__get_file__ok_200__nominal_case(self) -> None: """ Get one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file" ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "Test_file" assert content["parent_id"] == 1 assert content["show_in_ui"] is True assert content["slug"] == "test-file" assert content["status"] == "open" assert content["workspace_id"] == 1 assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p>description</p>" assert content["mimetype"] == "plain/text" assert content["size"] == len(b"Test file") assert content["file_extension"] == ".txt" assert content["filename"] == "Test_file.txt" assert content["page_nb"] == 1 assert content["has_pdf_preview"] is True assert content["has_jpeg_preview"] is True def test_api__get_file__ok_200__no_file_add(self) -> None: """ Get one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "Test file" assert content["parent_id"] == 1 assert content["show_in_ui"] is True assert content["slug"] == "test-file" assert content["status"] == "open" assert content["workspace_id"] == 1 assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "" assert content["mimetype"] == "" assert content["file_extension"] == "" assert content["filename"] == "Test file" assert content["size"] is None assert content["page_nb"] is None assert content["has_pdf_preview"] is False assert content["has_jpeg_preview"] is False def test_api__get_file__ok_200__binary_file(self) -> None: """ Get one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.bin", new_mimetype="application/octet-stream", new_content=bytes(100), ) content_api.save(test_file) dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "Test_file" assert content["parent_id"] == 1 assert content["show_in_ui"] is True assert content["slug"] == "test-file" assert content["status"] == "open" assert content["workspace_id"] == 1 assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "" assert content["mimetype"] == "application/octet-stream" assert content["size"] == 100 assert content["file_extension"] == ".bin" assert content["filename"] == "Test_file.bin" assert content["page_nb"] is None assert content["has_pdf_preview"] is False assert content["has_jpeg_preview"] is False def test_api__get_files__err_400__wrong_content_type(self) -> None: """ Get one file of a content content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/files/6", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_TYPE_NOT_ALLOWED def test_api__get_file__err_400__content_does_not_exist(self) -> None: """ Get one file (content 170 does not exist in db """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/1/files/170", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND def test_api__get_file__err_400__content_not_in_workspace(self) -> None: """ Get one file (content 9 is in workspace 2) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/1/files/9", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND def test_api__get_file__err_400__workspace_does_not_exist(self) -> None: """ Get one file (Workspace 40 does not exist) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/40/files/9", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.WORKSPACE_NOT_FOUND def test_api__get_file__err_400__workspace_id_is_not_int(self) -> None: """ Get one file, workspace id is not int """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/coucou/files/9", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.WORKSPACE_INVALID_ID def test_api__get_file__err_400__content_id_is_not_int(self) -> None: """ Get one file, content_id is not int """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/files/coucou", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_INVALID_ID def test_api__update_file_info_err_400__empty_label(self) -> None: """ Update(put) one file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file" ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400 ) # INFO - G.M - 2018-09-10 - Handle by marshmallow schema assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR def test_api__update_file_info__ok_200__nominal_case(self) -> None: """ Update(put) one file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) with dbsession.no_autoflush: test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) content_api.update_file_data( test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file" ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 1 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 1 assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["mimetype"] == "plain/text" assert content["size"] == len(b"Test file") assert content["page_nb"] == 1 assert content["has_pdf_preview"] is True assert content["has_jpeg_preview"] is True res = self.testapp.get( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 1 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 1 assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["mimetype"] == "plain/text" assert content["size"] == len(b"Test file") assert content["page_nb"] == 1 assert content["has_pdf_preview"] is True assert content["has_jpeg_preview"] is True def test_api__update_file_info__err_400__content_status_closed(self) -> None: """ Update(put) one file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) with dbsession.no_autoflush: test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) content_api.update_file_data( test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file" ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") test_file.status = "closed-validated" content_api.save(test_file) dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400 ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE def test_api__update_file_info__err_400__content_deleted(self) -> None: """ Update(put) one file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi( current_user=admin, session=dbsession, config=self.app_config, show_deleted=True ) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) with dbsession.no_autoflush: test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) content_api.update_file_data( test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file" ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") test_file.is_deleted = True content_api.save(test_file) dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400 ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE def test_api__update_file_info__err_400__content_archived(self) -> None: """ Update(put) one file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi( current_user=admin, session=dbsession, config=self.app_config, show_deleted=True ) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) with dbsession.no_autoflush: test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) content_api.update_file_data( test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file" ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") test_file.is_archived = True content_api.save(test_file) dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400 ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE def test_api__update_file_info__err_400__not_modified(self) -> None: """ Update(put) one file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file" ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 1 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 1 assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["mimetype"] == "plain/text" assert content["size"] == len(b"Test file") assert content["page_nb"] == 1 assert content["has_pdf_preview"] is True assert content["has_jpeg_preview"] is True res = self.testapp.get( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 1 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 1 assert content["current_revision_id"] # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["mimetype"] == "plain/text" assert content["size"] == len(b"Test file") assert content["page_nb"] == 1 assert content["has_pdf_preview"] is True assert content["has_jpeg_preview"] is True res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400 ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.SAME_VALUE_ERROR def test_api__update_file_info__err_400__label_already_used(self) -> None: """ Update(put) one file, failed because label already used """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="folder_used", do_save=True, do_notify=False, ) with dbsession.no_autoflush: test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") test_file2 = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, filename="already_used.txt", do_save=False, do_notify=False, ) test_file2.file_extension = ".txt" test_file2.depot_file = FileIntent(b"Test file", "already_used.txt", "text/plain") with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "folder_used", "raw_content": "<p> Le nouveau contenu </p>"} self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=200 ) params = {"label": "already_used", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400 ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER def test_api__get_file_revisions__ok_200__nominal_case(self) -> None: """ Get file revisions """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file" ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/1/files/{}/revisions".format(test_file.content_id), status=200 ) revisions = res.json_body assert len(revisions) == 1 revision = revisions[0] assert revision["content_type"] == "file" assert revision["content_id"] == test_file.content_id assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is True assert revision["label"] == "Test_file" assert revision["parent_id"] == 1 assert revision["show_in_ui"] is True assert revision["slug"] == "test-file" assert revision["status"] == "open" assert revision["workspace_id"] == 1 assert revision["revision_id"] assert revision["sub_content_types"] # TODO - G.M - 2018-06-173 - Test with real comments assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 1 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Global manager" assert revision["mimetype"] == "plain/text" assert revision["size"] == len(b"Test file") assert revision["page_nb"] == 1 assert revision["has_pdf_preview"] is True def test_api__set_file_status__ok_200__nominal_case(self) -> None: """ set file status """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "closed-deprecated"} # before res = self.testapp.get( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["status"] == "open" # set status self.testapp.put_json( "/api/v2/workspaces/1/files/{}/status".format(test_file.content_id), params=params, status=204, ) # after res = self.testapp.get( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["status"] == "closed-deprecated" def test_api__set_file_status__err_400__wrong_status(self) -> None: """ set file status """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "unexistant-status"} # before res = self.testapp.get( "/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200 ) content = res.json_body assert content["content_type"] == "file" assert content["content_id"] == test_file.content_id assert content["status"] == "open" # set status res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}/status".format(test_file.content_id), params=params, status=400, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR def test_api__get_file_raw__ok_200__nominal_case(self) -> None: """ Get one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) filename = "Test_file.txt" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, filename), status=200 ) assert res.body == b"Test file" assert res.content_type == "text/plain" assert res.content_length == len(b"Test file") assert int(res.headers["Content-Length"]) == res.content_length assert res.last_modified.second == test_file.updated.second assert res.last_modified.minute == test_file.updated.minute assert res.last_modified.day == test_file.updated.day assert res.last_modified.month == test_file.updated.month assert res.last_modified.year == test_file.updated.year def test_api__get_file_raw__ok_200__force_download_case(self) -> None: """ Get one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, new_content=b"Test file", new_filename="Test_file.txt", new_mimetype="text/plain", ) content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"force_download": 1} filename = "Test_file.txt" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, filename), status=200, params=params, ) assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename) assert res.body == b"Test file" assert res.content_type == "text/plain" assert res.content_length == len(b"Test file") assert int(res.headers["Content-Length"]) == res.content_length assert res.last_modified.second == test_file.updated.second assert res.last_modified.minute == test_file.updated.minute assert res.last_modified.day == test_file.updated.day assert res.last_modified.month == test_file.updated.month assert res.last_modified.year == test_file.updated.year def test_api__create_file__ok__200__nominal_case(self) -> None: """ create one file of a content at workspace root """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) image = create_1000px_png_test_image() res = self.testapp.post( "/api/v2/workspaces/{}/files".format(business_workspace.workspace_id), upload_files=[("files", image.name, image.getvalue())], status=200, ) res = res.json_body assert res["parent_id"] is None assert res["content_type"] == "file" assert res["is_archived"] is False assert res["is_deleted"] is False assert res["is_editable"] is True assert res["workspace_id"] == business_workspace.workspace_id assert isinstance(res["content_id"], int) content_id = res["content_id"] assert res["status"] == "open" assert res["label"] == "test_image" assert res["slug"] == "test-image" res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/files/{content_id}".format( workspace_id=business_workspace.workspace_id, content_id=content_id ), status=200, ) res = res.json_body assert res["parent_id"] is None assert res["content_type"] == "file" assert res["is_archived"] is False assert res["is_deleted"] is False assert res["is_editable"] is True assert res["workspace_id"] == business_workspace.workspace_id assert isinstance(res["content_id"], int) assert res["status"] == "open" assert res["label"] == "test_image" assert res["slug"] == "test-image" assert res["author"]["user_id"] == admin.user_id assert res["page_nb"] == 1 assert res["mimetype"] == "image/png" def test_api__create_file__err_400__filename_already_used(self) -> None: """ create one file of a content but filename is already used here """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) image = create_1000px_png_test_image() res = self.testapp.post( "/api/v2/workspaces/{}/files".format(business_workspace.workspace_id), upload_files=[("files", image.name, image.getvalue())], status=200, ) res = res.json_body assert res["parent_id"] is None assert res["content_type"] == "file" assert res["is_archived"] is False assert res["is_deleted"] is False assert res["is_editable"] is True assert res["workspace_id"] == business_workspace.workspace_id assert isinstance(res["content_id"], int) assert res["status"] == "open" assert res["label"] == "test_image" assert res["slug"] == "test-image" res = self.testapp.post( "/api/v2/workspaces/{}/files".format(business_workspace.workspace_id), upload_files=[("files", image.name, image.getvalue())], status=400, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER def test_api__create_file__ok__200__in_folder(self) -> None: """ create one file of a content in a folder """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) folder = content_api.create( label="test-folder", content_type_slug=content_type_list.Folder.slug, workspace=business_workspace, do_save=True, do_notify=False, ) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"parent_id": folder.content_id} image = create_1000px_png_test_image() res = self.testapp.post( "/api/v2/workspaces/{}/files".format(business_workspace.workspace_id), upload_files=[("files", image.name, image.getvalue())], params=params, status=200, ) res = res.json_body assert res["parent_id"] == folder.content_id assert res["content_type"] == "file" assert res["is_archived"] is False assert res["is_deleted"] is False assert res["is_editable"] is True assert res["workspace_id"] == business_workspace.workspace_id assert isinstance(res["content_id"], int) content_id = res["content_id"] assert res["status"] == "open" assert res["label"] == "test_image" assert res["slug"] == "test-image" res = self.testapp.get( "/api/v2/workspaces/{workspace_id}/files/{content_id}".format( workspace_id=business_workspace.workspace_id, content_id=content_id ), status=200, ) res = res.json_body assert res["parent_id"] == folder.content_id assert res["content_type"] == "file" assert res["is_archived"] is False assert res["is_deleted"] is False assert res["is_editable"] is True assert res["workspace_id"] == business_workspace.workspace_id assert isinstance(res["content_id"], int) assert res["status"] == "open" assert res["label"] == "test_image" assert res["slug"] == "test-image" assert res["author"]["user_id"] == admin.user_id assert res["page_nb"] == 1 assert res["mimetype"] == "image/png" def test_api__create_file__err__400__unallow_subcontent(self) -> None: """ create one file of a content but subcontent of type file unallowed here """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) folder = content_api.create( label="test-folder", content_type_slug=content_type_list.Folder.slug, workspace=business_workspace, do_save=True, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=folder): content_api.set_allowed_content(folder, []) content_api.save(folder) transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"parent_id": folder.content_id} image = create_1000px_png_test_image() res = self.testapp.post( "/api/v2/workspaces/{}/files".format(business_workspace.workspace_id), upload_files=[("files", image.name, image.getvalue())], params=params, status=400, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.UNALLOWED_SUBCONTENT def test_api__create_file__err__400__parent_not_found(self) -> None: """ create one file of a content but parent_id is not valid """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"parent_id": 3000} image = create_1000px_png_test_image() res = self.testapp.post( "/api/v2/workspaces/{}/files".format(business_workspace.workspace_id), upload_files=[("files", image.name, image.getvalue())], params=params, status=400, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.PARENT_NOT_FOUND def test_api__set_file_raw__ok_200__nominal_case(self) -> None: """ Set one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), status=200 ) assert res.body == image.getvalue() assert res.content_type == "image/png" assert res.content_length == len(image.getvalue()) def test_api__set_file_raw__ok_200__filename_already_used(self) -> None: """ Set one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file_2 = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file2", do_save=False, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) content2_id = int(test_file_2.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content2_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=400, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER def test_api__set_file_raw__err_400__closed_status_file(self) -> None: """ Set one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.status = "closed-validated" content_api.save(test_file) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=400, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE @pytest.mark.xfail(raises=AssertionError, reason="Broken feature dues to pyramid behaviour") def test_api__set_file_raw__err_400_not_modified(self) -> None: """ Set one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), status=200 ) assert res.body == image.getvalue() assert res.content_type == "image/png" assert res.content_length == len(image.getvalue()) res = self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status="*", ) assert res.status == 400 assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER def test_api__get_allowed_size_dim__ok__nominal_case(self) -> None: dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) content_id = int(test_file.content_id) res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/allowed_dims".format(content_id), status=200 ) res = res.json_body assert res["restricted"] is True assert len(res["dimensions"]) == 1 dim = res["dimensions"][0] assert dim["width"] == 256 assert dim["height"] == 256 def test_api__get_jpeg_preview__ok__200__nominal_case(self) -> None: """ Set one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/".format(content_id), status=200 ) assert res.body != image.getvalue() assert res.content_type == "image/jpeg" def test_api__get_jpeg_preview__ok__200__force_download_case(self) -> None: """ Set one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) params = {"force_download": 1} res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/raw".format(content_id), status=200, params=params, ) filename = "test_image_page_1.jpg" assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename) assert res.body != image.getvalue() assert res.content_type == "image/jpeg" def test_api__get_jpeg_preview__err_400__UnavailablePreview(self) -> None: """ Set one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.bin", new_mimetype="application/octet-stream", new_content=bytes(100), ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"force_download": 0} res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/".format(content_id), status=400, params=params, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.UNAIVALABLE_PREVIEW def test_api__get_sized_jpeg_preview__ok__200__nominal_case(self) -> None: """ get 256x256 preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/256x256/{}".format(content_id, image.name), status=200, ) assert res.body != image.getvalue() assert res.content_type == "image/jpeg" new_image = Image.open(io.BytesIO(res.body)) assert 256, 256 == new_image.size def test_api__get_sized_jpeg_preview__err_400__UnavailablePreview(self) -> None: """ Set one file of a content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.bin", new_mimetype="application/octet-stream", new_content=bytes(100), ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"force_download": 0} res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/256x256/{}".format( content_id, "Test_file.bin" ), status=400, params=params, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.UNAIVALABLE_PREVIEW def test_api__get_sized_jpeg_preview__ok__200__force_download_case(self) -> None: """ get 256x256 preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) params = {"force_download": 1} dl_filename = "test_image_page_1_256x256.jpg" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/256x256/{}".format(content_id, dl_filename), status=200, params=params, ) assert res.body != image.getvalue() assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(dl_filename, dl_filename) assert res.content_type == "image/jpeg" new_image = Image.open(io.BytesIO(res.body)) assert 256, 256 == new_image.size def test_api__get_sized_jpeg_preview__ok__200__force_download_case_no_filename(self) -> None: """ get 256x256 preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) params = {"force_download": 1} dl_filename = "test_image_page_1_256x256.jpg" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/256x256/".format(content_id), status=200, params=params, ) assert res.body != image.getvalue() assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(dl_filename, dl_filename) assert res.content_type == "image/jpeg" new_image = Image.open(io.BytesIO(res.body)) assert 256, 256 == new_image.size def test_api__get_sized_jpeg_preview__ok__200__force_download_case_filename_is_raw( self ) -> None: """ get 256x256 preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) params = {"force_download": 1} dl_filename = "test_image_page_1_256x256.jpg" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/256x256/raw".format(content_id), status=200, params=params, ) assert res.body != image.getvalue() assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(dl_filename, dl_filename) assert res.content_type == "image/jpeg" new_image = Image.open(io.BytesIO(res.body)) assert 256, 256 == new_image.size def test_api__get_sized_jpeg_preview__err__400__SizeNotAllowed(self) -> None: """ get 256x256 preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) filename = "test_image_512x512.jpg" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/jpg/512x512/{}".format(content_id, filename), status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.PREVIEW_DIM_NOT_ALLOWED def test_api__get_sized_jpeg_revision_preview__ok__200__nominal_case(self) -> None: """ get 256x256 revision preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) revision_id = int(test_file.revision_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) filename = "test_file.txt" res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format( content_id=content_id, revision_id=revision_id, filename=filename ), status=200, ) assert res.content_type == "text/plain" filename = "test_image_256x256.jpg" res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/jpg/256x256/{filename}".format( content_id=content_id, revision_id=revision_id, filename=filename ), status=200, ) assert res.body != image.getvalue() assert res.content_type == "image/jpeg" new_image = Image.open(io.BytesIO(res.body)) assert 256, 256 == new_image.size def test_api__get_sized_jpeg_revision_preview__ok__200__force_download_case(self) -> None: """ get 256x256 revision preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) revision_id = int(test_file.revision_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format( content_id=content_id, revision_id=revision_id, filename=image.name ), status=200, ) assert res.content_type == "text/plain" params = {"force_download": 1} res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/jpg/256x256/".format( content_id=content_id, revision_id=revision_id ), status=200, params=params, ) filename = "Test file_r{}_page_1_256x256.jpg".format(revision_id) urlencoded_filename = quote(filename) assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format( filename, urlencoded_filename ) assert res.body != image.getvalue() assert res.content_type == "image/jpeg" new_image = Image.open(io.BytesIO(res.body)) assert 256, 256 == new_image.size def test_api__get_full_pdf_preview__ok__200__nominal_case(self) -> None: """ get full pdf preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, test_file.file_name), upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())], status=204, ) filename = "test_image.pdf" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, filename), status=200, ) assert res.content_type == "application/pdf" def test_api__get_full_pdf_preview__ok__200__force_download_case(self) -> None: """ get full pdf preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) filename = "Test_file.txt" self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, filename), upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())], status=204, ) params = {"force_download": 1} res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, filename), status=200, params=params, ) assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename) assert res.content_type == "application/pdf" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, "Test_file.pdf"), status=200, params=params, ) filename = "Test_file.pdf" assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename) assert res.content_type == "application/pdf" def test_api__get_full_pdf_preview__err__400__png_UnavailablePreviewType(self) -> None: """ get full pdf preview of a png image -> error UnavailablePreviewType """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, image.name), status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.UNAVAILABLE_PREVIEW_TYPE def test_api__get_full_pdf_preview__err__400__png_UnavailablePreview(self) -> None: """ get full pdf preview of a png image -> error UnavailablePreviewType """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.bin", new_mimetype="application/octet-stream", new_content=bytes(100), ) dbsession.flush() content_id = test_file.content_id transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) filename = "Test_file.bin" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, filename), status=400, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.UNAIVALABLE_PREVIEW def test_api__get_pdf_preview__ok__200__nominal_case(self) -> None: """ get full pdf preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, test_file.file_name), upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())], status=204, ) params = {"page": 1} filename = "test_file.pdf" res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/{}".format(content_id, filename), status=200, params=params, ) assert res.content_type == "application/pdf" def test_api__get_pdf_preview_err__400__UnavailablePreview(self) -> None: """ get full pdf preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_file_data( test_file, "Test_file.bin", new_mimetype="application/octet-stream", new_content=bytes(100), ) dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"page": 1} res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/".format(content_id), status=400, params=params, ) assert isinstance(res.json, dict) assert "code" in res.json.keys() assert res.json_body["code"] == ErrorCode.UNAIVALABLE_PREVIEW def test_api__get_pdf_preview__ok__200__force_download_case(self) -> None: """ get full pdf preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) filename = "test_file.txt" self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, filename), upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())], status=204, ) filename = "Test_file_page_1.pdf" params = {"page": 1, "force_download": 1} res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/{}".format(content_id, filename), status=200, params=params, ) assert res.content_type == "application/pdf" assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename) def test_api__get_pdf_preview__ok__err__400_page_of_preview_not_found(self) -> None: """ get full pdf preview of a txt file """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=True, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_file): test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/".format(content_id), upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())], status=204, ) params = {"page": 2} res = self.testapp.get( "/api/v2/workspaces/1/files/{}/preview/pdf/".format(content_id), status=400, params=params, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.PAGE_OF_PREVIEW_NOT_FOUND def test_api__get_pdf_revision_preview__ok__200__nominal_case(self) -> None: """ get pdf revision preview of content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) revision_id = int(test_file.revision_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) filename = image.name res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format( content_id=content_id, revision_id=revision_id, filename=filename ), status=200, ) assert res.content_type == "text/plain" params = {"page": 1} filename = "test_image__page_1.pdf" res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/pdf/{filename}".format( content_id=content_id, revision_id=revision_id, params=params, filename=filename ), status=200, ) assert res.content_type == "application/pdf" def test_api__get_full_pdf_revision_preview__ok__200__nominal_case(self) -> None: """ get pdf revision preview of content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) revision_id = int(test_file.revision_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/".format( content_id=content_id, revision_id=revision_id ), status=200, ) assert res.content_type == "text/plain" res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/pdf/full/".format( content_id=content_id, revision_id=revision_id ), status=200, ) assert res.content_type == "application/pdf" def test_api__get_full_pdf_revision_preview__ok__200__force_download_case(self) -> None: """ get pdf revision preview of content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) revision_id = int(test_file.revision_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format( content_id=content_id, revision_id=revision_id, filename=image.name ), status=200, ) assert res.content_type == "text/plain" params = {"force_download": 1} filename = "Test_file.pdf" res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/pdf/full/{filename}".format( content_id=content_id, revision_id=revision_id, filename="Test_file.pdf" ), status=200, params=params, ) assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename) assert res.content_type == "application/pdf" def test_api__get_pdf_revision_preview__ok__200__force_download_case(self) -> None: """ get pdf revision preview of content """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") dbsession.flush() transaction.commit() content_id = int(test_file.content_id) revision_id = int(test_file.revision_id) image = create_1000px_png_test_image() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) self.testapp.put( "/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), upload_files=[("files", image.name, image.getvalue())], status=204, ) res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format( content_id=content_id, revision_id=revision_id, filename=image.name ), status=200, ) assert res.content_type == "text/plain" params = {"page": 1, "force_download": 1} filename = "test_image_page_1.pdf" res = self.testapp.get( "/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/pdf/{filename}".format( content_id=content_id, revision_id=revision_id, filename=filename ), status=200, params=params, ) assert res.headers[ "Content-Disposition" ] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename) assert res.content_type == "application/pdf" def test_api__set_file_status__err_400__same_status(self) -> None: dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_file = content_api.create( content_type_slug=content_type_list.File.slug, workspace=business_workspace, parent=tool_folder, label="Test file", do_save=False, do_notify=False, ) test_file.file_extension = ".txt" test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain") with new_revision(session=dbsession, tm=transaction.manager, content=test_file): content_api.update_content(test_file, "Test_file", "<p>description</p>") dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "open"} # set status res = self.testapp.put_json( "/api/v2/workspaces/1/files/{}/status".format(test_file.content_id), params=params, status=400, ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.INVALID_STATUS_CHANGE class TestThreads(FunctionalTest): """ Tests for /api/v2/workspaces/{workspace_id}/threads/{content_id} endpoint """ fixtures = [BaseFixture, ContentFixtures] def test_api__get_thread__err_400__wrong_content_type(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/threads/6", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_TYPE_NOT_ALLOWED def test_api__get_thread__ok_200__nominal_case(self) -> None: """ Get one html document of a content """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200) content = res.json_body assert content["content_type"] == "thread" assert content["content_id"] == 7 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "Best Cakes?" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "best-cakes" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 26 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] != content["author"] assert content["last_modifier"]["user_id"] == 3 assert content["last_modifier"]["public_name"] == "Bob i." assert content["last_modifier"]["avatar_url"] is None assert content["raw_content"] == "What is the best cake?" assert content["file_extension"] == ".thread.html" assert content["filename"] == "Best Cakes?.thread.html" def test_api__get_thread__err_400__content_does_not_exist(self) -> None: """ Get one thread (content 170 does not exist) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/threads/170", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND def test_api__get_thread__err_400__content_not_in_workspace(self) -> None: """ Get one thread(content 7 is in workspace 2) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/1/threads/7", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND def test_api__get_thread__err_400__workspace_does_not_exist(self) -> None: """ Get one thread (Workspace 40 does not exist) """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/40/threads/7", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.WORKSPACE_NOT_FOUND def test_api__get_thread__err_400__workspace_id_is_not_int(self) -> None: """ Get one thread, workspace id is not int """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/coucou/threads/7", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.WORKSPACE_INVALID_ID def test_api__get_thread__err_400_content_id_is_not_int(self) -> None: """ Get one thread, content id is not int """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/threads/coucou", status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.CONTENT_INVALID_ID def test_api__update_thread__ok_200__nominal_case(self) -> None: """ Update(put) thread """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json("/api/v2/workspaces/2/threads/7", params=params, status=200) content = res.json_body assert content["content_type"] == "thread" assert content["content_id"] == 7 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 28 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["file_extension"] == ".thread.html" assert content["filename"] == "My New label.thread.html" res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200) content = res.json_body assert content["content_type"] == "thread" assert content["content_id"] == 7 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 28 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" assert content["file_extension"] == ".thread.html" assert content["filename"] == "My New label.thread.html" def test_api__update_thread__err_400__not_modified(self) -> None: """ Update(put) thread """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json("/api/v2/workspaces/2/threads/7", params=params, status=200) content = res.json_body assert content["content_type"] == "thread" assert content["content_id"] == 7 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 28 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200) content = res.json_body assert content["content_type"] == "thread" assert content["content_id"] == 7 assert content["is_archived"] is False assert content["is_deleted"] is False assert content["is_editable"] is True assert content["label"] == "My New label" assert content["parent_id"] == 3 assert content["show_in_ui"] is True assert content["slug"] == "my-new-label" assert content["status"] == "open" assert content["workspace_id"] == 2 assert content["current_revision_id"] == 28 # TODO - G.M - 2018-06-173 - check date format assert content["created"] assert content["author"] assert content["author"]["user_id"] == 1 assert content["author"]["avatar_url"] is None assert content["author"]["public_name"] == "Global manager" # TODO - G.M - 2018-06-173 - check date format assert content["modified"] assert content["last_modifier"] == content["author"] assert content["raw_content"] == "<p> Le nouveau contenu </p>" res = self.testapp.put_json("/api/v2/workspaces/2/threads/7", params=params, status=400) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.SAME_VALUE_ERROR def test_api__update_thread__err_400__empty_label(self) -> None: """ Update(put) thread """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"label": "", "raw_content": "<p> Le nouveau contenu </p>"} res = self.testapp.put_json("/api/v2/workspaces/2/threads/7", params=params, status=400) # TODO - G.M - 2018-09-10 - Handle by marshmallow schema assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR def test_api__get_thread_revisions__ok_200__nominal_case(self) -> None: """ Get threads revisions """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get("/api/v2/workspaces/2/threads/7/revisions", status=200) revisions = res.json_body assert len(revisions) == 2 revision = revisions[0] assert revision["content_type"] == "thread" assert revision["content_id"] == 7 assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is False assert revision["label"] == "Best Cake" assert revision["parent_id"] == 3 assert revision["show_in_ui"] is True assert revision["slug"] == "best-cake" assert revision["status"] == "open" assert revision["workspace_id"] == 2 assert revision["revision_id"] == 8 assert revision["sub_content_types"] assert revision["revision_type"] == "creation" assert revision["comment_ids"] == [18, 19, 20] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 1 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Global manager" assert revision["file_extension"] == ".thread.html" assert revision["filename"] == "Best Cake.thread.html" revision = revisions[1] assert revision["content_type"] == "thread" assert revision["content_id"] == 7 assert revision["is_archived"] is False assert revision["is_deleted"] is False assert revision["is_editable"] is True assert revision["label"] == "Best Cakes?" assert revision["parent_id"] == 3 assert revision["show_in_ui"] is True assert revision["slug"] == "best-cakes" assert revision["status"] == "open" assert revision["workspace_id"] == 2 assert revision["revision_id"] == 26 assert revision["revision_type"] == "edition" assert revision["sub_content_types"] assert revision["comment_ids"] == [] # TODO - G.M - 2018-06-173 - check date format assert revision["created"] assert revision["author"] assert revision["author"]["user_id"] == 3 assert revision["author"]["avatar_url"] is None assert revision["author"]["public_name"] == "Bob i." assert revision["file_extension"] == ".thread.html" assert revision["filename"] == "Best Cakes?.thread.html" def test_api__get_thread_revisions__ok_200__most_revision_type(self) -> None: """ get threads revisions """ dbsession = get_tm_session(self.session_factory, transaction.manager) admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one() workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config) business_workspace = workspace_api.get_one(1) content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config) tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG) test_thread = content_api.create( content_type_slug=content_type_list.Thread.slug, workspace=business_workspace, parent=tool_folder, label="Test Thread", do_save=True, do_notify=False, ) with new_revision(session=dbsession, tm=transaction.manager, content=test_thread): content_api.update_content( test_thread, new_label="test_thread_updated", new_content="Just a test" ) content_api.save(test_thread) with new_revision(session=dbsession, tm=transaction.manager, content=test_thread): content_api.archive(test_thread) content_api.save(test_thread) with new_revision(session=dbsession, tm=transaction.manager, content=test_thread): content_api.unarchive(test_thread) content_api.save(test_thread) with new_revision(session=dbsession, tm=transaction.manager, content=test_thread): content_api.delete(test_thread) content_api.save(test_thread) with new_revision(session=dbsession, tm=transaction.manager, content=test_thread): content_api.undelete(test_thread) content_api.save(test_thread) dbsession.flush() transaction.commit() self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) res = self.testapp.get( "/api/v2/workspaces/1/threads/{}/revisions".format(test_thread.content_id), status=200 ) revisions = res.json_body assert len(revisions) == 6 for revision in revisions: assert revision["content_type"] == "thread" assert revision["workspace_id"] == 1 assert revision["content_id"] == test_thread.content_id revision = revisions[0] assert revision["revision_type"] == "creation" assert revision["is_editable"] is False revision = revisions[1] assert revision["revision_type"] == "edition" assert revision["is_editable"] is False revision = revisions[2] assert revision["revision_type"] == "archiving" assert revision["is_editable"] is False revision = revisions[3] assert revision["revision_type"] == "unarchiving" assert revision["is_editable"] is False revision = revisions[4] assert revision["revision_type"] == "deletion" assert revision["is_editable"] is False revision = revisions[5] assert revision["revision_type"] == "undeletion" assert revision["is_editable"] is True def test_api__set_thread_status__ok_200__nominal_case(self) -> None: """ Set thread status """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "closed-deprecated"} # before res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200) content = res.json_body assert content["content_type"] == "thread" assert content["content_id"] == 7 assert content["status"] == "open" assert content["is_editable"] is True # set status self.testapp.put_json("/api/v2/workspaces/2/threads/7/status", params=params, status=204) # after res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200) content = res.json_body assert content["content_type"] == "thread" assert content["content_id"] == 7 assert content["status"] == "closed-deprecated" assert content["is_editable"] is False def test_api__set_thread_status__ok_400__wrong_status(self) -> None: """ Set thread status """ self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "unexistant-status"} res = self.testapp.put_json( "/api/v2/workspaces/2/threads/7/status", params=params, status=400 ) # INFO - G.M - 2018-09-10 - Handle by marshmallow schema assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR def test_api__set_thread_status__err_400__same_status(self) -> None: self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin")) params = {"status": "open"} res = self.testapp.put_json( "/api/v2/workspaces/2/threads/7/status", params=params, status=400 ) assert res.json_body assert "code" in res.json_body assert res.json_body["code"] == ErrorCode.INVALID_STATUS_CHANGE
46.768891
116
0.631496
22,803
190,630
5.0396
0.014998
0.05778
0.064089
0.055344
0.978219
0.971884
0.967829
0.958231
0.947667
0.938869
0
0.017034
0.245517
190,630
4,075
117
46.780368
0.781967
0.040088
0
0.841667
0
0.003161
0.17451
0.039783
0
0
0
0.005644
0.327874
1
0.030172
false
0
0.005172
0
0.037644
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
7e47b4d7ceb33445c32cfe6565caf6173d824aa9
169
py
Python
textattack/attack_results/__init__.py
cclauss/TextAttack
98b8d6102aa47bf3c41afedace0215d48f8ed046
[ "MIT" ]
1
2021-06-24T19:35:18.000Z
2021-06-24T19:35:18.000Z
textattack/attack_results/__init__.py
53X/TextAttack
e6a7969abc1e28a2a8a7e2ace709b78eb9dc94be
[ "MIT" ]
null
null
null
textattack/attack_results/__init__.py
53X/TextAttack
e6a7969abc1e28a2a8a7e2ace709b78eb9dc94be
[ "MIT" ]
1
2021-11-12T05:26:21.000Z
2021-11-12T05:26:21.000Z
from .failed_attack_result import FailedAttackResult from .skipped_attack_result import SkippedAttackResult from .successful_attack_result import SuccessfulAttackResult
42.25
60
0.911243
18
169
8.222222
0.555556
0.243243
0.364865
0
0
0
0
0
0
0
0
0
0.071006
169
3
61
56.333333
0.942675
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
7e6895509707dd5ac5d1572800cdb46aafc739c4
2,352
py
Python
Evaluation/cluster.py
wangaxe/kmeans_defense
2dc0ac9aa7fea1bffcacdbf3f0f02694f3b6a328
[ "MIT" ]
null
null
null
Evaluation/cluster.py
wangaxe/kmeans_defense
2dc0ac9aa7fea1bffcacdbf3f0f02694f3b6a328
[ "MIT" ]
null
null
null
Evaluation/cluster.py
wangaxe/kmeans_defense
2dc0ac9aa7fea1bffcacdbf3f0f02694f3b6a328
[ "MIT" ]
null
null
null
from sklearn.cluster import KMeans, MiniBatchKMeans import numpy as np import torch def kmean_process(original_img, n_cluster_center): depth, height, width = original_img.shape pixel_sample = np.reshape(original_img, (height*width, depth)) compressed_img = np.zeros((height, width, depth), dtype=np.float32) kmeans = KMeans(n_clusters=n_cluster_center) kmeans.fit(pixel_sample) cluster_assignments = kmeans.predict(pixel_sample) cluster_centers = kmeans.cluster_centers_ pixel_count = 0 for i in range(height): for j in range(width): cluster_idx = cluster_assignments[pixel_count] cluster_value = cluster_centers[cluster_idx] compressed_img[i][j] = cluster_value pixel_count += 1 compressed_img = np.reshape(compressed_img, (depth, height, width)) return compressed_img def mini_bench_kmean_process(original_img, n_cluster_center, batch_size=100): depth, height, width = original_img.shape pixel_sample = np.reshape(original_img, (height*width, depth)) compressed_img = np.zeros((height, width, depth), dtype=np.float32) kmeans = MiniBatchKMeans(n_clusters=n_cluster_center, batch_size=batch_size) kmeans.fit(pixel_sample) cluster_assignments = kmeans.predict(pixel_sample) cluster_centers = kmeans.cluster_centers_ pixel_count = 0 for i in range(height): for j in range(width): cluster_idx = cluster_assignments[pixel_count] cluster_value = cluster_centers[cluster_idx] compressed_img[i][j] = cluster_value pixel_count += 1 compressed_img = np.reshape(compressed_img, (depth, height, width)) return compressed_img def Kmeans_cluster(in_tensor, k = 2): examples = in_tensor.detach().cpu().numpy() assert len(examples.shape) == 4 inter_res = np.array([kmean_process(example, k) for example in examples]) assert inter_res.shape == examples.shape res = torch.from_numpy(inter_res).cuda() return res def mb_Kmeans_cluster(in_tensor, k = 2): examples = in_tensor.detach().cpu().numpy() assert len(examples.shape) == 4 inter_res = np.array([mini_bench_kmean_process(example, k) for example in examples]) assert inter_res.shape == examples.shape res = torch.from_numpy(inter_res).cuda() return res
36.184615
88
0.711735
316
2,352
5.031646
0.202532
0.081761
0.03522
0.028931
0.918239
0.877987
0.877987
0.831447
0.831447
0.831447
0
0.007895
0.192177
2,352
64
89
36.75
0.828947
0
0
0.784314
0
0
0
0
0
0
0
0
0.078431
1
0.078431
false
0
0.058824
0
0.215686
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
7e811ebe6327b53430261c8c363fb3b3fb53953d
44,984
py
Python
py/genetargeter/recodeSelection.py
pablocarderam/genetargeter
4cd257d02e0753c1be9fa339aa505bb895315a6d
[ "MIT" ]
null
null
null
py/genetargeter/recodeSelection.py
pablocarderam/genetargeter
4cd257d02e0753c1be9fa339aa505bb895315a6d
[ "MIT" ]
2
2021-06-02T15:28:52.000Z
2021-06-17T12:08:53.000Z
py/genetargeter/recodeSelection.py
pablocarderam/genetargeter
4cd257d02e0753c1be9fa339aa505bb895315a6d
[ "MIT" ]
null
null
null
from builtins import str from builtins import range from py.utils.BioUtils import *; # Imports utils from py.utils.GenBankToolbox import *; # Imports utils from py.genetargeter.constants import *; # Imports constants from gRNAScores.gRNAScoring import *; # Import scoring metrics """ Chooses the region to be recoded to avoid gRNA targeting in already transfected regions. Returns GenBankAnn object with recoded sequence and indexes between which it should go. GenBank object given as argument should contain one gene with geneName included in its label, and at least one annotation with "LHR" in its label. Also needs all gRNAs to be annotated in the file. Returns empty region if LHR end is at or downstream of gene stop codon. Checks against restriction sites given as parameters. Checks that gRNA recoded sequence has a pairwise off-target score lower than the given threshold with respect to the original gRNA. """ def chooseRecodeRegion3Prime(geneGB, gene, offTargetMethod="cfd", pamType="NGG", orgCodonTable=codonUsage(), targetRegionOverride=False, filterCutSites=[cut_FseI,cut_AsiSI,cut_IPpoI,cut_ISceI,cut_AflII,cut_AhdI,cut_BsiWI,cut_NheI], codonSampling=False, offScoreThreshold=10, minGCEnd=0.375, gRNATableString=""): #TODO: debug #TODO: Recoded if upstream of stop codon add recode values to table gRNAs = geneGB.findAnnsLabel("gRNA", True); # List of all gRNAs gRNATable = gRNATableString.split('\n'); # split string into lines gRNATable = [g.split(',') for g in gRNATable]; # split each line into values if offTargetMethod == "hsu": # if off-target scoring with Hsu scores offScoreThreshold = 1; # set threshold to 1% log = ""; # init log LHRs = geneGB.findAnnsLabel("LHR"); # LHR annotation objects RHRs = geneGB.findAnnsLabel("RHR"); # LHR annotation objects LHR = None RHR = None for ann in LHRs: if '(unused)' not in ann.label: LHR = ann break for ann in RHRs: if '(unused)' not in ann.label: RHR = ann break annRecoded = GenBankAnn(); # creates GenBankAnn object to hold recoded region if LHR.index[1] < gene.index[1]: # if end of LHR is inside gene (or before) startRecode = max(LHR.index[1], gene.index[0]); # start of recode region (end of LHR or start of gene, most downstream) while not geneGB.checkInExon(startRecode) and startRecode <= len(geneGB.origin): # while recode region start is in intron, startRecode += 1; # shift downstream intronStartIndices = []; # stores start indexes of introns starting after recode sequence start intronEndIndices = []; # stores end indexes of introns starting after recode sequence start for ann in geneGB.findAnnsLabel(gene.label): # loop through annotations associated with transcript if ann.type == "CDS": # if annotation is cds if gene.index[1] > ann.index[1] > startRecode: # if annotation is an exon ending before gene end and after recode start, intronStartIndices.append(ann.index[1]); # add this intron start index if gene.index[1] > ann.index[0] > startRecode: # if annotation is an exon starting after recode start, intronEndIndices.append(ann.index[0]); # add this intron end index # if len(intronStartIndices) == 0: # if no CDS exons found in this way, # for ann in geneGB.findAnnsLabel(gene.label.split('.')[0]): # loop through annotations associated with gene # if ann.type == "exon": # if annotation is exon # if gene.index[1] > ann.index[1] > startRecode: # if annotation is an exon ending before gene end and after recode start, # intronStartIndices.append(ann.index[1]); # add this intron start index # if gene.index[1] > ann.index[0] > startRecode: # if annotation is an exon starting after recode start, # intronEndIndices.append(ann.index[0]); # add this intron end index # # # elif ann.type == "intron": # if annotation is intron, # if ann.index[0] > startRecode: # if annotation is an intron starting after recode start, # intronStartIndices.append(ann.index[0]); # add this intron start index # intronEndIndices.append(ann.index[1]); # add this intron start index intronIndices = []; # will contain final indexes of introns downstream of recode start (introns to be removed from recoded region) intronStartIndices = sorted(intronStartIndices); # sort intronEndIndices = sorted(intronEndIndices); # sort for i in range(len(intronEndIndices)): # for every intron end, if intronEndIndices[i] >= LHR.index[1]: # if after LHR end, start = 0; # will store corresponding intron start (largest underneath intron end) for startIndex in intronStartIndices: # loop through starts if startIndex > intronEndIndices[i]: # if start surpasses this end, break; # stop loop else: # if not, start = startIndex; # set as start intronIndices.append([start,intronEndIndices[i]]); # add these coordinates to intron splice list #for g in gRNAs: # for every gRNA, #if not (intronStartIndices[i] >= g.index[1] or intronEndIndices[i] <= g.index[0]): # if intron and gRNA overlap, #intronIndices.append([intronStartIndices[i],intronEndIndices[i]]); # add these coordinates to intron splice list endRecode = min(gene.index[1],RHR.index[0]) if targetRegionOverride else gene.index[1] - 3; # end of recode region (end of gene, exclude stop codon) recodeSeq = geneGB.origin[startRecode:endRecode]; # will contain sequence to be recorded nonRecodedEnd = '' frame2 = 0 if len(intronIndices) > 0 and intronIndices[0][0] < endRecode: # if there are introns inside the target region, recodeSeq = geneGB.origin[startRecode:intronIndices[0][0]]; # get recode sequence until first intron for i in range(len(intronIndices)-1): # for every intron except last one, if intronIndices[i][1] < endRecode: recodeSeq = recodeSeq + geneGB.origin[intronIndices[i][1]:min(intronIndices[i+1][0],endRecode)]; # add next exon to recode seq if intronIndices[len(intronIndices)-1][1] < endRecode: recodeSeq = recodeSeq + geneGB.origin[intronIndices[len(intronIndices)-1][1]:endRecode]; # get rest of recode sequence until endRecode # Adjust frame if not recoding to the stop codon if targetRegionOverride: restSeq = geneGB.origin[startRecode:intronIndices[0][0]]; # get recode sequence until first intron for i in range(len(intronIndices)-1): # for every intron except last one, restSeq = restSeq + geneGB.origin[intronIndices[i][1]:intronIndices[i+1][0]]; # add next exon to recode seq restSeq = restSeq + geneGB.origin[intronIndices[len(intronIndices)-1][1]:gene.index[1]]; # get rest of recode sequence until endRecode frame2 = 3-((len(restSeq)-len(recodeSeq)) % 3); # stores reading frame, index from start of sequence to be recoded frame2 = frame2 if frame2 != 3 else 0 endRecode -= frame2; # modify recode start site according to reading frame nonRecodedEnd = recodeSeq[-frame2:] if frame2!=0 else ''; # stores 0, 1 or 2 nucleotides not recoded due to reading frame recodeSeq = recodeSeq[0:len(recodeSeq)-frame2]; # adjust recode region frame = len(recodeSeq) % 3; # stores reading frame, index from start of sequence to be recoded startRecode += frame; # modify recode start site according to reading frame nonRecodedStart = recodeSeq[0:frame]; # stores 0, 1 or 2 nucleotides not recoded due to reading frame recodeSeq = recodeSeq[frame:len(recodeSeq)]; # adjust recode region cutSeqs = filterCutSites + [g.seq for g in gRNAs]; # list of all cut seqs. all gRNAs in gene are to be included as cut sequences cutCheck = True; # variable used to check if no cut sequences are present. offScore = 100; # stores off-target score. Default is 100% due to the fact that gRNA sequence is the same. count = 0; # iteration counter recodedSeq = recodeSeq; # assign recoded sequence to same as original bestRecodedSeq = recodedSeq; # will store best candidate sequence if len(recodeSeq) > 2: # if recodeSeq contains at least one codon, tricky = -1; # > -1 if suspected to be hard to synthesize badStart = False; # True if first bases have low melting temp (important for Gibson assembly) candidateFound = False; # signal possible candidate found bestRecodedSeq = recodedSeq; # will store best candidate sequence while not cutCheck or offScore > offScoreThreshold or tricky > -1 or badStart: # while cutCheck shows hits in a cut sequences, or while the pairwise off-target score is over the threshold, or while there are difficult-to-synthesize structures in the recoded region, or while the first 40 bp have a bad gc content if count == 1: # if recoded region has failed checks once, log = log + "Defaulted recoded region recodonization to codon sampling due to possible difficulties in synthesis or enzyme cut sites.\n\n"; # log warning codonSampling = True; # forces codonSampling to true if count == 10: log = log + "Defaulted recoded region recodonization to random codon sampling due to possible difficulties in synthesis or enzyme cut sites.\n\n"; # log warning orgCodonTable = codonUsage(); # forces random codon selection to true cutCheck = True; # reset cutCheck offScore = 0; # reset offScore tricky = -1; # reset tricky index badStart = False; # reset badStart Boolean recodedSeq = optimizeCodons(recodeSeq,orgCodonTable,codonSampling=codonSampling); # optimize codons. for g in gRNAs: # for every gRNA candidate within recoded region, if g.index[0] >= startRecode-frame and g.index[1] <= endRecode+frame2: # if grna is inside recoded region gOnSeq = g.seq; # get original gRNA sequence wholeRecSeq = nonRecodedStart + recodedSeq; # add initial bases gOffSeq = ""; anchor = -1; # will store index of gRNA bp most to the left (whichever strand). Default to -1 to indicate excision if geneGB.checkInExon(g.index[0]) or geneGB.checkInExon(g.index[1]): # if the gRNA hasn't been completely excised, if pamType == "NGG" and g.comp or pamType == "TTTV" and not g.comp: # if PAM is to the left of the rest of the gRNA sequence (on whichever strand), anchor = g.index[0]-startRecode-frame; # stores index of gRNA bp most to the left (whichever strand) for intron in intronIndices: # for every intron, if g.index[0] > intron[1]: # if anchor after end of intron, anchor -= intron[1]-intron[0]; # substract intron length from anchor index elif intron[0] >= g.index[0] >= intron[1]: # if anchor inside intron, anchor -= g.index[0] - intron[0]; # substract distance between intron start and anchor from anchor gOffSeq = wholeRecSeq[anchor:anchor+len(g.seq)]; # get recoded sequence that used to be gRNA if g.comp: # if on comp strand gOffSeq = revComp(gOffSeq); # save as reverse complement else: # if PAM is to the right, anchor = g.index[1]-startRecode-frame; # stores index of gRNA bp most to the right (whichever strand) for intron in intronIndices: # for every intron, if g.index[1] > intron[1]: # if anchor after end of intron, anchor -= intron[1]-intron[0]; # substract intron length from anchor index elif intron[0] >= g.index[1] >= intron[1]: # if anchor inside intron, anchor -= g.index[1] - intron[0]; # substract distance between intron start and anchor from anchor gOffSeq = wholeRecSeq[anchor-len(g.seq):anchor]; # get recoded sequence that used to be gRNA if g.comp: # if on comp strand gOffSeq = revComp(gOffSeq); # save as reverse complement gNewPAM = ""; # will store new PAM sequence if pamType == "NGG" and anchor > -1: # if using NGG PAM and gRNA not completely excised, if (g.index[1]+3 >= endRecode and not g.comp) or (g.index[0]-3 >= startRecode and g.comp): # if PAM is within recoded region, if not g.comp: # if on positive strand, gNewPAM = wholeRecSeq[anchor+len(g.seq):anchor+len(g.seq)+3]; # retrieve PAM downstream of gRNA sequence else: # if on negative strand, gNewPAM = revComp(wholeRecSeq[anchor+len(g.seq)-3:anchor+len(g.seq)]); # retrieve PAM upstream of gRNA sequence, on comp strand else: # if outside recoded region, if g.comp: # if on comp strand, gNewPAM = geneGB.origin[g.index[1]:g.index[1]+3]; # will store new PAM sequence else: # if on positive strand, gNewPAM = revComp(geneGB.origin[g.index[0]-3:g.index[0]]); # will store new PAM sequence elif pamType == "TTTV" and anchor > -1: # if using TTTV PAM and gRNA not completely excised, if (g.index[1]+4 >= endRecode and g.comp) or (g.index[0]-4 >= startRecode and not g.comp): # if PAM is inside recoded region, if not g.comp: # if on positive strand, gNewPAM = wholeRecSeq[anchor+len(g.seq)-4:anchor+len(g.seq)]; # retrieve PAM upstream of gRNA sequence else: # if on negative strand, gNewPAM = revComp(wholeRecSeq[anchor+len(g.seq):anchor+len(g.seq)+4]); # retrieve PAM downstream of gRNA sequence, on comp strand else: # if outside recoded region, if g.comp: # if on comp strand, gNewPAM = geneGB.origin[g.index[1]:g.index[1]+4]; # will store new PAM sequence else: # if on positive strand, gNewPAM = revComp(geneGB.origin[g.index[0]-4:g.index[0]]); # will store new PAM sequence newOffScore = 0; # Assume gRNA was excised if offTargetMethod == "cfd" and len(gOffSeq) > 22: # if using cfd and gRNA not completely excised, newOffScore = pairScoreCFD(gOnSeq,gOffSeq,gNewPAM,pamType); # calculate pairwise off-target score elif offTargetMethod == "hsu" and len(gOffSeq) > 22: # if using hsu and gRNA not completely excised, newOffScore = pairScoreHsu(gOnSeq,gOffSeq,gNewPAM,pamType); # calculate pairwise off-target score offScore = max(offScore,newOffScore); # set offscore for next iteration for g in gRNATable: # find this gRNA in table if "gRNAs not evaluated" not in gRNATableString and g[14] == gOnSeq: # if there is a gRNA table (no table if using custom gRNA) and gRNA found, g[15] = gOffSeq; # store recoded sequence g[16] = str(newOffScore); # store recoded sequence's pair score else: # if gRNA is not entirely contained, offScore = max(offScore,0); # assume recoded for site in cutSeqs: # for every cut site being filtered, cutCheck = cutCheck * ( findFirst(recodedSeq,site) < 0 ); # Find cut site, register in cutCheck cutCheck = cutCheck * ( findFirst(recodedSeq,revComp(site)) < 0 ); # Find cut site in comp strand, register in cutCheck if gcContent(recodedSeq[0:40]) < minGCEnd: # if the first bases don't have enough gc content badStart = True; trickyCount = 1 trickyLimit = 1000 tricky = isTricky(recodedSeq); # check if tricky to synthesize bestRecodedSeq = recodedSeq if bestRecodedSeq==recodeSeq else bestRecodedSeq; # store this sequence if no recoded sequence has been stored as best if offScore <= offScoreThreshold and cutCheck: # if parameters other than badStart are ok and this sequence has better start than previous best, if not candidateFound or isTricky(bestRecodedSeq) > -1: # if no candidate found until now or current best is already tricky, while tricky > -1 and tricky < len(recodedSeq)-9 and trickyCount < trickyLimit: # targeted recoding of problematic fragments recodedSeq = recodedSeq[0:tricky-tricky%3] + optimizeCodons(recodedSeq[tricky-tricky%3:tricky-tricky%3+9]) + recodedSeq[tricky-tricky%3+9:]; # optimize codons. new_tricky = isTricky(recodedSeq) tricky = max(tricky,new_tricky) if new_tricky > -1 else new_tricky; # check if tricky to synthesize (only downstream to avoid going back to fix newly repeated sequences) trickyCount += 1 if trickyCount % 10 == 0: # shuffle everything every 100 targeted recodings recodedSeq = recodedSeq[0:tricky-tricky%3] + optimizeCodons(recodedSeq[tricky-tricky%3:]); # optimize codons of remainder new_tricky = isTricky(recodedSeq) tricky = max(tricky,new_tricky) if new_tricky > -1 else new_tricky; # check if tricky to synthesize (only downstream to avoid going back to fix newly repeated sequences) bestRecodedSeq = recodedSeq; # make this new best elif not tricky > -1 and gcContent(recodedSeq[0:40]) > gcContent(bestRecodedSeq[0:40]): bestRecodedSeq = recodedSeq; # make this new best if not tricky > -1: candidateFound = True; # signal possible candidate found count += 1; # advances iteration counter if count > 200 or trickyCount >= trickyLimit: # if out of iteration limit, if not candidateFound: # if no candidate without cut sequences found, if tricky > -1: log = log + "Warning: Recoded region for gene " + gene.label + " could not reshuffle enough to avoid repeated sequences or low-complexity regions.\n\n"; # log warning else: log = log + "Warning: Recoded region for gene " + gene.label + " could not reshuffle enough to fulfill the maximum off-target sgRNA score threshold, or avoid all the following cut sequences: \n" + str(cutSeqs) + "\n\n"; # log warning break; # escape loop #print [gOnSeq+"NGG",gOffSeq+gNewPAM,pairScoreCFD(gOnSeq,gOffSeq,gNewPAM,pamType),pairScoreHsu(gOnSeq,gOffSeq,gNewPAM,pamType)] recodedSeq = nonRecodedStart + bestRecodedSeq + nonRecodedEnd; # adds initial bases from reading frame adjustment to best candidate annRecoded = GenBankAnn(gene.label + " Recoded", "misc_feature", recodedSeq, False, [startRecode-frame,endRecode+frame2], annColors['recodedRegionColor']); # creates var to store finished recodedSeq as annotation log = log + "Recoded region with size " + str(len(recodedSeq)) + " for gene " + gene.label + " selected.\n\n"; # logs this process finished else: # if no recoded region necessary, log = log + "Recoded region not deemed necessary for gene " + gene.label + ".\n\n"; # logs this process finished if "gRNAs not evaluated" not in gRNATableString: gRNATableString = "\n".join([",".join(g) for g in gRNATable]); # Creates string from grna array gRNATableString = gRNATableString.replace(">=threshold",">="+str(offScoreThreshold)); # adds pairwise recoded threshold values return {"out":annRecoded, "log":log, "gRNATable":gRNATableString}; # returns recoded region GenBankAnn object """ Chooses the region to be recoded to avoid gRNA targeting in already transfected regions. Returns GenBankAnn object with recoded sequence and indexes between which it should go. GenBank object given as argument should contain one gene with geneName included in its label, and at least one annotation with "LHR" in its label. Also needs all gRNAs to be annotated in the file. Returns empty region if LHR end is at or downstream of gene stop codon. Checks against restriction sites given as parameters. Checks that gRNA recoded sequence has a pairwise off-target score lower than the given threshold with respect to the original gRNA. """ def chooseRecodeRegion5Prime(geneGB, gene, offTargetMethod="cfd", pamType="NGG", orgCodonTable=codonUsage(), targetRegionOverride=False, filterCutSites=[cut_FseI,cut_AsiSI,cut_IPpoI,cut_ISceI,cut_AflII,cut_AhdI,cut_BsiWI,cut_NheI], codonSampling=False, offScoreThreshold=10, minGCEnd=0.375, gRNATableString="", haTag=True): #TODO: debug #TODO: Recoded if upstream of stop codon add recode values to table gRNAs = geneGB.findAnnsLabel("gRNA", True); # List of all gRNAs gRNATable = gRNATableString.split('\n'); # split string into lines gRNATable = [g.split(',') for g in gRNATable]; # split each line into values if offTargetMethod == "hsu": # if off-target scoring with Hsu scores offScoreThreshold = 1; # set threshold to 1% log = ""; # init log LHR = geneGB.findAnnsLabel("LHR")[0]; # LHR annotation object RHR = geneGB.findAnnsLabel("RHR")[0]; # RHR annotation object annRecoded = GenBankAnn(); # creates GenBankAnn object to hold recoded region if RHR.index[0] > gene.index[0]: # if end of RHR is inside gene endRecode = min(RHR.index[0],gene.index[1]); # end of recode region (start of RHR or end of gene, most upstream) while not geneGB.checkInExon(endRecode): # while recode region end is in intron, endRecode -= 1; # shift upstream intronStartIndices = []; # stores start indexes of introns starting after recode sequence start intronEndIndices = []; # stores end indexes of introns starting after recode sequence start for ann in geneGB.findAnnsLabel(gene.label): # loop through annotations associated with transcript if ann.type == "CDS": # if annotation is cds if gene.index[0] < ann.index[1] < endRecode: # if annotation is an exon ending after gene start and before recode end, intronStartIndices.append(ann.index[1]); # add this intron start index if gene.index[0] < ann.index[0] < endRecode: # if annotation is an exon starting before recode end, intronEndIndices.append(ann.index[0]); # add this intron end index # if len(intronStartIndices) == 0: # if no CDS exons found in this way, # for ann in geneGB.findAnnsLabel(gene.label.split('.')[0]): # loop through annotations associated with gene # if ann.type == "exon": # if annotation is exon # if gene.index[0] < ann.index[1] < endRecode: # if annotation is an exon ending after gene start and before recode end, # intronStartIndices.append(ann.index[1]); # add this intron start index # if gene.index[0] < ann.index[0] < endRecode: # if annotation is an exon starting before recode end, # intronEndIndices.append(ann.index[0]); # add this intron end index # # # elif ann.type == "intron": # if annotation is intron, # if ann.index[1] < endRecode: # if annotation is an intron ending before recode end, # intronStartIndices.append(ann.index[0]); # add this intron start index # intronEndIndices.append(ann.index[1]); # add this intron start index intronIndices = []; # will contain final indexes of introns downstream of recode start (introns to be removed from recoded region) intronStartIndices = sorted(intronStartIndices); # sort intronEndIndices = sorted(intronEndIndices); # sort for i in range(len(intronEndIndices)): # for every intron end, if intronEndIndices[i] <= RHR.index[0]: # if before RHR start, start = 0; # will store corresponding intron start (largest underneath intron end) for startIndex in intronStartIndices: # loop through starts if startIndex > intronEndIndices[i]: # if start surpasses this end, break; # stop loop else: # if not, start = startIndex; # set as start intronIndices.append([start,intronEndIndices[i]]); # add these coordinates to intron splice list #for g in gRNAs: # for every gRNA, #if not (intronStartIndices[i] >= g.index[1] or intronEndIndices[i] <= g.index[0]): # if intron and gRNA overlap, #intronIndices.append([intronStartIndices[i],intronEndIndices[i]]); # add these coordinates to intron splice list startRecode = max(gene.index[0],LHR.index[1]) if targetRegionOverride else gene.index[0]; # end of recode region (end of gene, exclude stop codon) recodeSeq = geneGB.origin[startRecode:endRecode]; # will contain sequence to be recorded nonRecodedStart = '' frame2 = 0 if len(intronIndices) > 0 and intronIndices[0][0] < endRecode: # if there are introns inside the target region, recodeSeq = geneGB.origin[startRecode:intronIndices[0][0]]; # get recode sequence until first intron for i in range(len(intronIndices)-1): # for every intron except last one, if intronIndices[i][1] < endRecode: recodeSeq = recodeSeq + geneGB.origin[intronIndices[i][1]:min(intronIndices[i+1][0],endRecode)]; # add next exon to recode seq if intronIndices[len(intronIndices)-1][1] < endRecode: recodeSeq = recodeSeq + geneGB.origin[intronIndices[len(intronIndices)-1][1]:endRecode]; # get rest of recode sequence until endRecode # Adjust frame if not recoding from the start codon if targetRegionOverride: restSeq = geneGB.origin[startRecode:intronIndices[0][0]]; # get recode sequence until first intron for i in range(len(intronIndices)-1): # for every intron except last one, restSeq = restSeq + geneGB.origin[intronIndices[i][1]:intronIndices[i+1][0]]; # add next exon to recode seq restSeq = restSeq + geneGB.origin[intronIndices[len(intronIndices)-1][1]:gene.index[1]]; # get rest of recode sequence until endRecode frame2 = len(recodeSeq) % 3; # stores reading frame, index from start of sequence to be recoded frame = 3-((len(restSeq)-len(recodeSeq)) % 3) # stores reading frame, index from start of sequence to be recoded frame = frame if frame != 3 else 0 startRecode += frame2; # modify recode start site according to reading frame nonRecodedStart = recodeSeq[0:frame2] if frame2!=0 else ''; # stores 0, 1 or 2 nucleotides not recoded due to reading frame recodeSeq = recodeSeq[frame2:]; # adjust recode region else: frame = len(recodeSeq) % 3; # stores reading frame, index from start of sequence to be recoded # frame = len(recodeSeq) % 3; # stores reading frame, index from start of sequence to be recoded endRecode -= frame; # modify recode end site according to reading frame nonRecodedEnd = ""; # stores 0, 1 or 2 nucleotides not recoded due to reading frame if frame != 0: # if frame shift is not zero, to avoid listing all of recode region with recodeSeq[-0:], nonRecodedEnd = recodeSeq[-frame:]; # stores 0, 1 or 2 nucleotides not recoded due to reading frame recodeSeq = recodeSeq[0:len(recodeSeq)-frame]; # adjust recode region if haTag: # if adding an HA tag, recodeSeq = ha_tag + recodeSeq; # add HA tag to start of recoded region cutSeqs = filterCutSites + [g.seq for g in gRNAs]; # list of all cut seqs. all gRNAs in gene are to be included as cut sequences cutCheck = True; # variable used to check if cut sequences are present. Initially false since all gRNAs are present. offScore = 100; # stores off-target score. Default is 100% due to the fact that gRNA sequence is the same. count = 0; # iteration counter recodedSeq = recodeSeq; # assign recoded sequence to same as original bestRecodedSeq = recodedSeq; # will store best candidate sequence if len(recodeSeq) > 2: # if recodeSeq contains at least one codon, tricky = -1; # True if suspected to be hard to synthesize badStart = False; # True if first bases have low melting temp (important for Gibson assembly) candidateFound = False; # signal possible candidate found bestRecodedSeq = recodedSeq; # will store best candidate sequence while not cutCheck or offScore > offScoreThreshold or tricky > -1 or badStart: # while cutCheck is greater than what you would expect for no hits in all cut sequences plus the gRNAs on both positive and comp strands, or while the pairwise off-target score is over the threshold, or while there are difficult-to-synthesize structures in the recoded region, or while the first 40 bp have a bad gc content if count == 1: # if recoded region has failed checks once, log = log + "Defaulted recoded region recodonization to codon sampling due to possible difficulties in synthesis or enzyme cut sites.\n\n"; # log warning codonSampling = True; # forces codonSampling to true if count == 10: log = log + "Defaulted recoded region recodonization to random codon sampling due to possible difficulties in synthesis or enzyme cut sites.\n\n"; # log warning orgCodonTable = codonUsage(); # forces random codon selection to true cutCheck = True; # reset cutCheck offScore = 0; # reset offScore tricky = -1; # reset tricky index badStart = False; # reset badStart Boolean recodedSeq = optimizeCodons(recodeSeq,orgCodonTable,codonSampling=codonSampling); # optimize codons. for g in gRNAs: # for every gRNA candidate within recoded region, if g.index[0] >= startRecode-frame2 and g.index[1] <= endRecode+frame: # if grna is inside recoded region gOnSeq = g.seq; # get original gRNA sequence wholeRecSeq = nonRecodedStart + recodedSeq + nonRecodedEnd; # add initial bases gOffSeq = ""; anchor = -1; # will store index of gRNA bp most to the left (whichever strand). Default to -1 to indicate excision if geneGB.checkInExon(g.index[0]) or geneGB.checkInExon(g.index[1]): # if the gRNA hasn't been completely excised, if pamType == "NGG" and g.comp or pamType == "TTTV" and not g.comp: # if PAM is to the left of the rest of the gRNA sequence (on whichever strand), anchor = g.index[0]-startRecode-frame2; # stores index of gRNA bp most to the left (whichever strand) for intron in intronIndices: # for every intron, if g.index[0] > intron[1]: # if anchor after end of intron, anchor -= intron[1]-intron[0]; # substract intron length from anchor index elif intron[0] >= g.index[0] >= intron[1]: # if anchor inside intron, anchor -= g.index[0] - intron[0]; # substract distance between intron start and anchor from anchor gOffSeq = wholeRecSeq[anchor:anchor+len(g.seq)]; # get recoded sequence that used to be gRNA if g.comp: # if on comp strand gOffSeq = revComp(gOffSeq); # save as reverse complement else: # if PAM is to the right, anchor = g.index[1]-startRecode-frame2; # stores index of gRNA bp most to the right (whichever strand) for intron in intronIndices: # for every intron, if g.index[1] > intron[1]: # if anchor after end of intron, anchor -= intron[1]-intron[0]; # substract intron length from anchor index elif intron[0] >= g.index[1] >= intron[1]: # if anchor inside intron, anchor -= g.index[1] - intron[0]; # substract distance between intron start and anchor from anchor gOffSeq = wholeRecSeq[anchor-len(g.seq):anchor]; # get recoded sequence that used to be gRNA if g.comp: # if on comp strand gOffSeq = revComp(gOffSeq); # save as reverse complement gNewPAM = ""; # will store new PAM sequence if pamType == "NGG" and anchor > -1: # if using NGG PAM and gRNA not completely excised, if (g.index[1]+3 >= endRecode and not g.comp) or (g.index[0]-3 >= startRecode and g.comp): # if PAM is within recoded region, if not g.comp: # if on positive strand, gNewPAM = wholeRecSeq[anchor+len(g.seq):anchor+len(g.seq)+3]; # retrieve PAM downstream of gRNA sequence else: # if on negative strand, gNewPAM = revComp(wholeRecSeq[anchor+len(g.seq)-3:anchor+len(g.seq)]); # retrieve PAM upstream of gRNA sequence, on comp strand else: # if outside recoded region, if g.comp: # if on comp strand, gNewPAM = geneGB.origin[g.index[1]:g.index[1]+3]; # will store new PAM sequence else: # if on positive strand, gNewPAM = revComp(geneGB.origin[g.index[0]-3:g.index[0]]); # will store new PAM sequence elif pamType == "TTTV" and anchor > -1: # if using TTTV PAM and gRNA not completely excised, if (g.index[1]+4 >= endRecode and g.comp) or (g.index[0]-4 >= startRecode and not g.comp): # if PAM is inside recoded region, if not g.comp: # if on positive strand, gNewPAM = wholeRecSeq[anchor+len(g.seq)-4:anchor+len(g.seq)]; # retrieve PAM upstream of gRNA sequence else: # if on negative strand, gNewPAM = revComp(wholeRecSeq[anchor+len(g.seq):anchor+len(g.seq)+4]); # retrieve PAM downstream of gRNA sequence, on comp strand else: # if outside recoded region, if g.comp: # if on comp strand, gNewPAM = geneGB.origin[g.index[1]:g.index[1]+4]; # will store new PAM sequence else: # if on positive strand, gNewPAM = revComp(geneGB.origin[g.index[0]-4:g.index[0]]); # will store new PAM sequence newOffScore = 0; # Assume gRNA was excised if offTargetMethod == "cfd" and len(gOffSeq) > 22: # if using cfd and gRNA not completely excised, newOffScore = pairScoreCFD(gOnSeq,gOffSeq,gNewPAM,pamType); # calculate pairwise off-target score elif offTargetMethod == "hsu" and len(gOffSeq) > 22: # if using hsu and gRNA not completely excised, newOffScore = pairScoreHsu(gOnSeq,gOffSeq,gNewPAM,pamType); # calculate pairwise off-target score offScore = max(offScore,newOffScore); # set offscore for next iteration for g in gRNATable: # find this gRNA in table if "gRNAs not evaluated" not in gRNATableString and g[14] == gOnSeq: # if there is a gRNA table (no table if using custom gRNA) and gRNA found, g[15] = gOffSeq; # store recoded sequence g[16] = str(newOffScore); # store recoded sequence's pair score else: # if gRNA is not entirely contained, offScore = max(offScore,0); # assume recoded for site in cutSeqs: # for every cut site being filtered, cutCheck = cutCheck * ( findFirst(recodedSeq,site) < 0 ); # Find cut site, register in cutCheck cutCheck = cutCheck * ( findFirst(recodedSeq,revComp(site)) < 0 ); # Find cut site in comp strand, register in cutCheck if gcContent(recodedSeq[-40:]) < minGCEnd: # if the last bases don't have enough gc content badStart = True; trickyCount = 1 trickyLimit = 1000 tricky = isTricky(recodedSeq); # check if tricky to synthesize bestRecodedSeq = recodedSeq if bestRecodedSeq==recodeSeq else bestRecodedSeq; # store this sequence if no recoded sequence has been stored as best if offScore <= offScoreThreshold and cutCheck: # if parameters other than badStart are ok and this sequence has better start than previous best, if not candidateFound or isTricky(bestRecodedSeq) > -1: # if no candidate found until now or current best is already tricky, while tricky > -1 and tricky < len(recodedSeq)-9 and trickyCount < trickyLimit: # targeted recoding of problematic fragments recodedSeq = recodedSeq[0:tricky-tricky%3] + optimizeCodons(recodedSeq[tricky-tricky%3:tricky-tricky%3+9]) + recodedSeq[tricky-tricky%3+9:]; # optimize codons. new_tricky = isTricky(recodedSeq) tricky = max(tricky,new_tricky) if new_tricky > -1 else new_tricky; # check if tricky to synthesize (only downstream to avoid going back to fix newly repeated sequences) trickyCount += 1 if trickyCount % 10 == 0: # shuffle everything every 100 targeted recodings recodedSeq = recodedSeq[0:tricky-tricky%3] + optimizeCodons(recodedSeq[tricky-tricky%3:]); # optimize codons of remainder new_tricky = isTricky(recodedSeq) tricky = max(tricky,new_tricky) if new_tricky > -1 else new_tricky; # check if tricky to synthesize (only downstream to avoid going back to fix newly repeated sequences) bestRecodedSeq = recodedSeq; # make this new best elif not tricky > -1 and gcContent(recodedSeq[-40:]) > gcContent(bestRecodedSeq[-40:]): bestRecodedSeq = recodedSeq; # make this new best if not tricky > -1: candidateFound = True; # signal possible candidate found count += 1; # advances iteration counter if count > 200 or trickyCount >= trickyLimit: # if out of iteration limit, if not candidateFound: # if no candidate without cut sequences found, if tricky > -1: log = log + "Warning: Recoded region for gene " + gene.label + " could not reshuffle enough to avoid repeated sequences or low-complexity regions.\n\n"; # log warning else: log = log + "Warning: Recoded region for gene " + gene.label + " could not reshuffle enough to fulfill the maximum off-target sgRNA score threshold, or avoid all the following cut sequences: \n" + str(cutSeqs) + "\n\n"; # log warning break; # escape loop #print [gOnSeq+"NGG",gOffSeq+gNewPAM,pairScoreCFD(gOnSeq,gOffSeq,gNewPAM,pamType),pairScoreHsu(gOnSeq,gOffSeq,gNewPAM,pamType)] recodedSeq = nonRecodedStart + bestRecodedSeq + nonRecodedEnd; # adds end bases from reading frame adjustment to best candidate annRecoded = GenBankAnn(gene.label + " Recoded", "misc_feature", recodedSeq, False, [startRecode-frame2,endRecode+frame], annColors['recodedRegionColor']); # creates var to store finished recodedSeq as annotation haTagMsg = ""; # used to output message if haTag: # if using an HA tag, haTagMsg = " with a recoded HA tag"; # msg modifier log = log + "Recoded region with size " + str(len(recodedSeq)) + " for gene " + gene.label + haTagMsg + " selected.\n\n"; # logs this process finished else: # if no recoded region necessary, log = log + "Recoded region not deemed necessary for gene " + gene.label + ".\n\n"; # logs this process finished if "gRNAs not evaluated" not in gRNATableString: gRNATableString = "\n".join([",".join(g) for g in gRNATable]); # Creates string from grna array gRNATableString = gRNATableString.replace(">=threshold",">="+str(offScoreThreshold)); # adds pairwise recoded threshold values return {"out":annRecoded, "log":log, "gRNATable":gRNATableString}; # returns recoded region GenBankAnn object """ Chooses the region to be recoded to avoid gRNA targeting in already transfected regions. Returns GenBankAnn object with recoded sequence and indexes between which it should go. GenBank object given as argument should contain one gene with geneName included in its label, and at least one annotation with "LHR" in its label. Also needs all gRNAs to be annotated in the file. Returns empty region if LHR end is at or downstream of gene stop codon. Checks against restriction sites given as parameters. Checks that gRNA recoded sequence has a pairwise off-target score lower than the given threshold with respect to the original gRNA. """ def chooseRecodeRegion(geneGB, gene, offTargetMethod="cfd", pamType="NGG", orgCodonTable=codonUsage(), targetRegionOverride=False, filterCutSites=[cut_FseI,cut_AsiSI,cut_IPpoI,cut_ISceI,cut_AflII,cut_AhdI,cut_BsiWI,cut_NheI], codonSampling=False, offScoreThreshold=10, minGCEnd=0.375, gRNATableString="", target3Prime=True, haTag=False): out = {}; # will contain method output if target3Prime: # if targeting 3' out = chooseRecodeRegion3Prime(geneGB, gene, offTargetMethod, pamType=pamType, orgCodonTable=orgCodonTable,codonSampling=codonSampling, gRNATableString=gRNATableString, targetRegionOverride=targetRegionOverride, filterCutSites=filterCutSites); # defines region to be recoded, returns recoded sequence else: # if using pSN150, out = chooseRecodeRegion5Prime(geneGB, gene, offTargetMethod, pamType=pamType, orgCodonTable=orgCodonTable,codonSampling=codonSampling, gRNATableString=gRNATableString, haTag=haTag, targetRegionOverride=targetRegionOverride, filterCutSites=filterCutSites); # defines region to be recoded, returns recoded sequence return out;
73.744262
414
0.618042
5,418
44,984
5.123477
0.083056
0.01124
0.006556
0.009366
0.927879
0.916424
0.913001
0.911884
0.909939
0.907922
0
0.014241
0.302219
44,984
609
415
73.865353
0.870114
0.369131
0
0.80198
0
0.014851
0.06244
0
0
0
0
0.003284
0
1
0.007426
false
0
0.014851
0
0.029703
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
7e8a6ba7ba5315f5b907de2d512c604d0dcf914f
88
py
Python
pytestJune2020/code/fixtures/test_add_pytest.py
Parazyne/my_presentations
0f4298761c05208ac2ab8f907cbd0471c60ad988
[ "MIT" ]
1
2020-06-19T03:15:22.000Z
2020-06-19T03:15:22.000Z
pytestJune2020/code/simple/test_add_pytest.py
Parazyne/my_presentations
0f4298761c05208ac2ab8f907cbd0471c60ad988
[ "MIT" ]
null
null
null
pytestJune2020/code/simple/test_add_pytest.py
Parazyne/my_presentations
0f4298761c05208ac2ab8f907cbd0471c60ad988
[ "MIT" ]
null
null
null
from add import add_two_numbers def test_add(): assert add_two_numbers(1, 3) == 4
14.666667
37
0.715909
16
88
3.625
0.6875
0.206897
0.448276
0
0
0
0
0
0
0
0
0.042254
0.193182
88
5
38
17.6
0.774648
0
0
0
0
0
0
0
0
0
0
0
0.333333
1
0.333333
true
0
0.333333
0
0.666667
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
0
1
0
1
0
0
8
0e8368a3facf91ee997929f09d55d3a1d854d4b4
158
py
Python
tests/cases/resources/tests/__init__.py
chop-dbhi/varify
5dc721e49ed9bd3582f4b117785fdd1a8b6ba777
[ "BSD-2-Clause" ]
6
2015-01-16T14:35:29.000Z
2017-06-18T05:56:15.000Z
tests/cases/resources/tests/__init__.py
solvebio/varify
5dc721e49ed9bd3582f4b117785fdd1a8b6ba777
[ "BSD-2-Clause" ]
null
null
null
tests/cases/resources/tests/__init__.py
solvebio/varify
5dc721e49ed9bd3582f4b117785fdd1a8b6ba777
[ "BSD-2-Clause" ]
3
2015-05-27T15:03:17.000Z
2020-03-11T08:42:46.000Z
from gene_rank import * # noqa from assessments import * # noqa from genes import * # noqa from phenotypes import * # noqa from variants import * # noqa
26.333333
33
0.71519
21
158
5.333333
0.428571
0.446429
0.5
0
0
0
0
0
0
0
0
0
0.221519
158
5
34
31.6
0.910569
0.151899
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
0e8f062a5146956f40087807fffcef101744fede
2,560
py
Python
tests/test_http_exception.py
VincentHch/fastapi-jsonrpc
aeba440322c30224bc121f5a5ce77ab354841989
[ "MIT" ]
155
2019-08-09T18:33:19.000Z
2022-03-23T11:32:44.000Z
tests/test_http_exception.py
VincentHch/fastapi-jsonrpc
aeba440322c30224bc121f5a5ce77ab354841989
[ "MIT" ]
29
2020-01-16T11:34:08.000Z
2022-02-16T15:23:57.000Z
tests/test_http_exception.py
VincentHch/fastapi-jsonrpc
aeba440322c30224bc121f5a5ce77ab354841989
[ "MIT" ]
24
2019-10-11T15:15:40.000Z
2022-02-17T11:54:21.000Z
import contextlib from json import dumps as json_dumps from fastapi import HTTPException from fastapi_jsonrpc import JsonRpcContext def test_method(ep, raw_request): @ep.method() def probe() -> str: raise HTTPException(401) resp = raw_request(json_dumps({ 'id': 1, 'jsonrpc': '2.0', 'method': 'probe', 'params': {}, })) assert resp.status_code == 401 assert resp.json() == {'detail': 'Unauthorized'} def test_ep_middleware_enter(ep, raw_request): @contextlib.asynccontextmanager async def middleware(_ctx: JsonRpcContext): raise HTTPException(401) # noinspection PyUnreachableCode yield ep.middlewares.append(middleware) @ep.method() def probe() -> str: return 'qwe' resp = raw_request(json_dumps({ 'id': 1, 'jsonrpc': '2.0', 'method': 'probe', 'params': {}, })) assert resp.status_code == 401 assert resp.json() == {'detail': 'Unauthorized'} def test_ep_middleware_exit(ep, raw_request): @contextlib.asynccontextmanager async def middleware(_ctx: JsonRpcContext): yield raise HTTPException(401) ep.middlewares.append(middleware) @ep.method() def probe() -> str: return 'qwe' resp = raw_request(json_dumps({ 'id': 1, 'jsonrpc': '2.0', 'method': 'probe', 'params': {}, })) assert resp.status_code == 401 assert resp.json() == {'detail': 'Unauthorized'} def test_method_middleware_enter(ep, raw_request): @contextlib.asynccontextmanager async def middleware(_ctx: JsonRpcContext): raise HTTPException(401) # noinspection PyUnreachableCode yield @ep.method(middlewares=[middleware]) def probe() -> str: return 'qwe' resp = raw_request(json_dumps({ 'id': 1, 'jsonrpc': '2.0', 'method': 'probe', 'params': {}, })) assert resp.status_code == 401 assert resp.json() == {'detail': 'Unauthorized'} def test_method_middleware_exit(ep, raw_request): @contextlib.asynccontextmanager async def middleware(_ctx: JsonRpcContext): yield raise HTTPException(401) @ep.method(middlewares=[middleware]) def probe() -> str: return 'qwe' resp = raw_request(json_dumps({ 'id': 1, 'jsonrpc': '2.0', 'method': 'probe', 'params': {}, })) assert resp.status_code == 401 assert resp.json() == {'detail': 'Unauthorized'}
22.654867
52
0.598828
267
2,560
5.595506
0.164794
0.066934
0.040161
0.060241
0.892905
0.880187
0.880187
0.880187
0.880187
0.880187
0
0.023923
0.265234
2,560
112
53
22.857143
0.770335
0.023828
0
0.890244
0
0
0.098958
0
0
0
0
0
0.121951
1
0.121951
false
0
0.04878
0.04878
0.219512
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
0eb15ce5e2d074f96d46fea515a175efade9aeb4
9,220
py
Python
icons/check_boxes.py
robert-hh/SSD1963-TFT-Library-for-PyBoard
db9786cdd95f9dab5334a9de28bed4e26436815c
[ "MIT" ]
16
2016-02-23T12:20:36.000Z
2021-02-02T06:41:49.000Z
icons/check_boxes.py
robert-hh/SSD1963-TFT-Library-for-PyBoard-and-RP2040
db9786cdd95f9dab5334a9de28bed4e26436815c
[ "MIT" ]
2
2016-11-26T07:46:58.000Z
2017-12-10T08:44:38.000Z
icons/check_boxes.py
robert-hh/SSD1963-TFT-Library-for-PyBoard
db9786cdd95f9dab5334a9de28bed4e26436815c
[ "MIT" ]
9
2016-06-04T08:22:55.000Z
2020-04-19T14:40:36.000Z
# Code generated by bmp_to_icon.py from uctypes import addressof _icons = { 0: ( b'\xff\xff\xff\xff\xff\xff\xff\xd7\xff\xff\xff\xff\xff\xff\xff\xd5' b'\xff\xff\xff\xff\xff\xff\xff\x55\xc0\x00\x00\x00\x00\x00\xfd\x57' b'\x00\x00\x00\x00\x00\x03\xf5\x57\x02\xaa\xaa\xaa\xaa\xaa\xa5\x5f' b'\x02\xaa\xaa\xaa\xaa\xaa\x95\x7f\x02\xaa\xaa\xaa\xaa\xaa\x55\xff' b'\x02\xaa\xaa\xaa\xaa\xaa\x55\xff\x02\xaa\xaa\xaa\xaa\xa9\x57\xcf' b'\x02\xaa\xaa\xaa\xaa\xa5\x5b\x0f\x02\xaa\xaa\xaa\xaa\xa5\x58\x0f' b'\x02\xaa\xaa\xaa\xaa\x95\x68\x0f\x02\xaa\xaa\xaa\xaa\x55\xa8\x0f' b'\x02\xaa\x5a\xaa\xaa\x55\xa8\x0f\x02\xa9\x56\xaa\xa9\x56\xa8\x0f' b'\x02\xaa\x55\xaa\xa5\x5a\xa8\x0f\x02\xaa\x55\x6a\x95\x5a\xa8\x0f' b'\x02\xaa\x95\x5a\x95\x6a\xa8\x0f\x02\xaa\xa5\x56\x55\xaa\xa8\x0f' b'\x02\xaa\xa9\x55\x56\xaa\xa8\x0f\x02\xaa\xaa\x55\x56\xaa\xa8\x0f' b'\x02\xaa\xaa\x95\x5a\xaa\xa8\x0f\x02\xaa\xaa\xa5\x6a\xaa\xa8\x0f' b'\x02\xaa\xaa\xa9\x6a\xaa\xa8\x0f\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f' b'\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f' b'\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f' b'\x00\x00\x00\x00\x00\x00\x00\x3f\xc0\x00\x00\x00\x00\x00\x00\xff' ), 1: ( b'\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff' b'\xff\xff\xff\xff\xff\xff\xff\xff\xc0\x00\x00\x00\x00\x00\x00\xff' b'\x00\x00\x00\x00\x00\x00\x00\x3f\x02\xaa\xaa\xaa\xaa\xaa\xa0\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x02\xaa\xaa\xaa\xaa\xaa\xa0\x3f' b'\x00\x00\x00\x00\x00\x00\x00\x3f\xc0\x00\x00\x00\x00\x00\x00\xff' ), 2: ( b'\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff' b'\xff\xff\xff\xff\xff\xff\xff\xff\xc0\x00\x00\x00\x00\x00\x00\xff' b'\x00\x00\x00\x00\x00\x00\x00\x3f\x02\xaa\xaa\xaa\xaa\xaa\xa0\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xa5\x55\x56\xaa\xa8\x3f' b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\x95\x55\x55\xaa\xa8\x3f' b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\x95\x55\x55\xaa\xa8\x3f' b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\x95\x55\x55\xaa\xa8\x3f' b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\x95\x55\x55\xaa\xa8\x3f' b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\xa5\x55\x56\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f' b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x02\xaa\xaa\xaa\xaa\xaa\xa0\x3f' b'\x00\x00\x00\x00\x00\x00\x00\x3f\xc0\x00\x00\x00\x00\x00\x00\xff' ), 3: ( b'\xff\xff\xff\x00\x00\xff\xff\xff\xff\xff\xc0\x00\x00\x03\xff\xff' b'\xff\xfc\x00\xaa\xaa\x00\x3f\xff\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff' b'\xff\xc0\xaa\xaa\xaa\xaa\x03\xff\xff\x02\xaa\xaa\xaa\xaa\x80\xff' b'\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f' b'\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f\xc0\xaa\xaa\xaa\xaa\xa6\xaa\x03' b'\xc2\xaa\xaa\xaa\xaa\x95\xaa\x83\xc2\xaa\xaa\xaa\xaa\x56\xaa\x83' b'\x0a\xaa\xaa\xaa\xa9\x56\xaa\x80\x0a\xaa\xaa\xaa\xa9\x5a\xaa\xa0' b'\x0a\xaa\xaa\xaa\xa5\x6a\xaa\xa0\x0a\xaa\xaa\xaa\x95\x6a\xaa\xa0' b'\x0a\xaa\xaa\xaa\x95\xaa\xaa\xa0\x0a\xaa\x96\xaa\x55\xaa\xaa\xa0' b'\x0a\xaa\x95\xaa\x56\xaa\xaa\xa0\x0a\xaa\x95\x69\x5a\xaa\xaa\x80' b'\xc2\xaa\xa5\x55\x5a\xaa\xaa\x83\xc2\xaa\xa9\x55\x6a\xaa\xaa\x83' b'\xc0\xaa\xaa\x95\xaa\xaa\xaa\x03\xf0\xaa\xaa\xa5\xaa\xaa\xaa\x0f' b'\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f' b'\xff\x02\xaa\xaa\xaa\xaa\x80\xff\xff\xc0\xaa\xaa\xaa\xaa\x03\xff' b'\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff\xff\xfc\x00\x2a\xa8\x00\x3f\xff' b'\xff\xff\xc0\x00\x00\x03\xff\xff\xff\xff\xff\x00\x00\xff\xff\xff' ), 4: ( b'\xff\xff\xff\x00\x00\xff\xff\xff\xff\xff\xc0\x00\x00\x03\xff\xff' b'\xff\xfc\x00\xaa\xaa\x00\x3f\xff\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff' b'\xff\xc0\xaa\xaa\xaa\xaa\x03\xff\xff\x02\xaa\xaa\xaa\xaa\x80\xff' b'\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f' b'\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03' b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83' b'\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0' b'\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0' b'\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0' b'\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0' b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83' b'\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f' b'\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f' b'\xff\x02\xaa\xaa\xaa\xaa\x80\xff\xff\xc0\xaa\xaa\xaa\xaa\x03\xff' b'\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff\xff\xfc\x00\xaa\xaa\x00\x3f\xff' b'\xff\xff\xc0\x00\x00\x03\xff\xff\xff\xff\xff\x00\x00\xff\xff\xff' ), 5: ( b'\xff\xff\xff\x00\x00\xff\xff\xff\xff\xff\xc0\x00\x00\x03\xff\xff' b'\xff\xfc\x00\xaa\xaa\x00\x3f\xff\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff' b'\xff\xc0\xaa\xaa\xaa\xaa\x03\xff\xff\x02\xaa\xaa\xaa\xaa\x80\xff' b'\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f' b'\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03' b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83' b'\x0a\xaa\xaa\xa5\x6a\xaa\xaa\xa0\x0a\xaa\xaa\x95\x5a\xaa\xaa\xa0' b'\x0a\xaa\xaa\x55\x56\xaa\xaa\xa0\x0a\xaa\xaa\x55\x56\xaa\xaa\xa0' b'\x0a\xaa\xaa\x55\x56\xaa\xaa\xa0\x0a\xaa\xaa\x95\x5a\xaa\xaa\xa0' b'\x0a\xaa\xaa\xa5\x6a\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0' b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83' b'\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f' b'\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f' b'\xff\x02\xaa\xaa\xaa\xaa\x80\xff\xff\xc0\xaa\xaa\xaa\xaa\x03\xff' b'\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff\xff\xfc\x00\xaa\xaa\x00\x3f\xff' b'\xff\xff\xc0\x00\x00\x03\xff\xff\xff\xff\xff\x00\x00\xff\xff\xff' ), 6: ( b'\xff\xff\xff\xfd\x7f\xff\xff\xff\xff\xff\xc0\xf5\x5f\x03\xff\xff' b'\xff\xfc\x00\xf5\x5f\x00\x3f\xff\xff\xf0\x0a\xa5\x5a\xa0\x0f\xff' b'\xff\xc0\xaa\xa5\x5a\xaa\x03\xff\xff\x02\xaa\xa5\x5a\xaa\x80\xff' b'\xfc\x0a\xaa\xa5\x5a\xaa\xa0\x3f\xf0\x2a\xaa\xa5\x5a\xaa\xa8\x0f' b'\xf0\xaa\xaa\xa5\x5a\xaa\xaa\x0f\xc0\xaa\xaa\xa5\x5a\xaa\xaa\x03' b'\xc2\xaa\xaa\xa5\x5a\xaa\xaa\x83\xc2\xaa\xaa\xa5\x5a\xaa\xaa\x83' b'\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0' b'\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0' b'\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0' b'\x0a\xaa\xaa\xa9\x6a\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0' b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83' b'\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f' b'\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f' b'\xff\x02\xaa\xaa\xaa\xaa\x80\xff\xff\xc0\xaa\xaa\xaa\xaa\x03\xff' b'\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff\xff\xfc\x00\xaa\xaa\x00\x3f\xff' b'\xff\xff\xc0\x00\x00\x03\xff\xff\xff\xff\xff\x00\x00\xff\xff\xff' ), } colortable = { 0: (b'\x00\x00\x00\x02\x00\x00\x00\x00\xff\xff\xff\x00\xff\xff\xff\x00'), 1: (b'\x00\x00\x00\x02\x01\x24\xf7\x00\xf0\xf0\xf0\x00\xff\xff\xff\x00'), 2: (b'\x00\x00\x00\x02\x00\xff\x00\x00\xf0\xf0\xf0\x00\xff\xff\xff\x00'), 3: (b'\x00\x00\x00\x02\x00\xce\xff\x00\xf0\xf0\xf0\x00\xff\xff\xff\x00'), 4: (b'\xff\xff\xff\x02\xff\xff\xff\x00\x00\x00\x00\x00\x00\x00\x00\x00'), 5: (b'\xff\xff\xff\x02\x01\x24\xf7\x00\x10\x10\x10\x00\x00\x00\x00\x00'), 6: (b'\xff\xff\xff\x02\x00\xff\x00\x00\x10\x10\x10\x00\x00\x00\x00\x00'), 7: (b'\xff\xff\xff\x02\x00\xce\xff\x00\x10\x10\x10\x00\x00\x00\x00\x00'), } width = 32 height = 32 colors = 2 def get_icon(icon_index = 0, color_index = 0): return width, height, addressof(_icons[icon_index]), colors, addressof(colortable[color_index]) def draw(x, y, icon_index, draw_fct, color_index = 0): draw_fct(x - width//2, y - height // 2, width, height, addressof(_icons[icon_index]), colors, addressof(colortable[color_index]))
60.261438
133
0.67603
2,125
9,220
2.925647
0.036706
0.512466
0.502332
0.434293
0.924883
0.878398
0.809394
0.760817
0.729452
0.703394
0
0.148496
0.07679
9,220
152
134
60.657895
0.581884
0.003471
0
0.575342
1
0.821918
0.836146
0.836146
0
1
0
0
0
1
0.013699
false
0
0.006849
0.006849
0.027397
0
0
0
0
null
1
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
1
1
1
1
null
1
0
0
0
0
0
0
0
0
0
0
0
0
14
7eba828afd143c2c50ba8c130be49d1be5cef779
3,737
py
Python
tests/test_binder.py
erayon/cycli
2947ad90396091652e3f43a9f61a74539e3f310e
[ "MIT" ]
290
2015-07-31T07:07:11.000Z
2021-12-15T09:39:59.000Z
tests/test_binder.py
erayon/cycli
2947ad90396091652e3f43a9f61a74539e3f310e
[ "MIT" ]
56
2015-08-01T14:47:15.000Z
2019-05-14T11:43:03.000Z
tests/test_binder.py
erayon/cycli
2947ad90396091652e3f43a9f61a74539e3f310e
[ "MIT" ]
40
2015-08-07T13:30:40.000Z
2021-05-02T00:58:09.000Z
from __future__ import unicode_literals from mock import call, MagicMock import pytest from cycli.binder import * @pytest.fixture def binder(): return MagicMock() def test_curly_left(binder): curly_left(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call("{"), call("}", move_cursor=False)]) def test_curly_right_curly_current(binder): binder.cli.current_buffer.document.current_char = "}" curly_right(binder) binder.cli.current_buffer.cursor_right.assert_called_once_with() def test_curly_right_not_curly_current(binder): binder.cli.current_buffer.document.current_char = "n" curly_right(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call("}")]) def test_paren_left(binder): paren_left(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call("("), call(")", move_cursor=False)]) def test_paren_right_paren_current(binder): binder.cli.current_buffer.document.current_char = ")" paren_right(binder) binder.cli.current_buffer.cursor_right.assert_called_once_with() def test_paren_right_not_paren_current(binder): binder.cli.current_buffer.document.current_char = "n" paren_right(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call(")")]) def test_bracket_left(binder): bracket_left(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call("["), call("]", move_cursor=False)]) def test_bracket_right_bracket_current(binder): binder.cli.current_buffer.document.current_char = "]" bracket_right(binder) binder.cli.current_buffer.cursor_right.assert_called_once_with() def test_bracket_right_not_bracket_current(binder): binder.cli.current_buffer.document.current_char = "n" bracket_right(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call("]")]) def test_apostrophe_current(binder): binder.cli.current_buffer.document.current_char = "'" apostrophe(binder) binder.cli.current_buffer.cursor_right.assert_called_once_with() def test_apostrophe_not_current(binder): binder.cli.current_buffer.document.current_char = "n" apostrophe(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call("'"), call("'", move_cursor=False)]) def test_quote_current(binder): binder.cli.current_buffer.document.current_char = "\"" quote(binder) binder.cli.current_buffer.cursor_right.assert_called_once_with() def test_quote_not_current(binder): binder.cli.current_buffer.document.current_char = "n" quote(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call("\""), call("\"", move_cursor=False)]) def test_backtick_current(binder): binder.cli.current_buffer.document.current_char = "`" backtick(binder) binder.cli.current_buffer.cursor_right.assert_called_once_with() def test_backtick_not_current(binder): binder.cli.current_buffer.document.current_char = "n" backtick(binder) binder.cli.current_buffer.insert_text.assert_has_calls([call("`"), call("`", move_cursor=False)]) def test_backspace_no_pattern(binder): binder.cli.current_buffer.document.current_char = "n" backspace(binder) binder.cli.current_buffer.delete_before_cursor.assert_called_once_with() @pytest.mark.parametrize(("char_before_cursor", "current_char"), [ ("(", ")"), ("[", "]"), ("{", "}"), ("'", "'"), ('"', '"'), ("`", "`") ]) def test_backspace_with_pattern(char_before_cursor, current_char): binder = MagicMock() binder.cli.current_buffer.document.char_before_cursor = char_before_cursor binder.cli.current_buffer.document.current_char = current_char backspace(binder) binder.cli.current_buffer.cursor_right.assert_called_once_with() binder.cli.current_buffer.delete_before_cursor.assert_called_once_with(2)
29.896
101
0.779235
509
3,737
5.333988
0.096267
0.109392
0.194475
0.267403
0.836832
0.805893
0.780111
0.752486
0.752486
0.675138
0
0.000293
0.086433
3,737
124
102
30.137097
0.794962
0
0
0.337349
0
0
0.020605
0
0
0
0
0
0.216867
1
0.216867
false
0
0.048193
0.012048
0.277108
0
0
0
0
null
0
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
0
0
0
0
0
9
7ed019aa7c9507a8c3e573da8a623a8d1f824c03
6,651
py
Python
Abt-Buy/rltk_exp/feature_vector.py
cybergla/rltk-experimentation
30485068c557bd218f40cc7db1d05cb3e1e61642
[ "MIT" ]
1
2018-08-15T17:58:42.000Z
2018-08-15T17:58:42.000Z
Abt-Buy/rltk_exp/feature_vector.py
cybergla/rltk-experimentation
30485068c557bd218f40cc7db1d05cb3e1e61642
[ "MIT" ]
null
null
null
Abt-Buy/rltk_exp/feature_vector.py
cybergla/rltk-experimentation
30485068c557bd218f40cc7db1d05cb3e1e61642
[ "MIT" ]
5
2018-10-26T16:02:45.000Z
2021-07-15T10:54:31.000Z
from create_datasets import * # def generate_feature_vector(r_abt, r_buy): # brand_score = 0.5 # if r_abt.brand_cleaned and r_buy.brand_cleaned: # if r_abt.brand_cleaned == r_buy.brand_cleaned: # brand_score = 1 # else: # if len(r_abt.brand_cleaned) >= len(r_buy.brand_cleaned): # common_str = r_buy.brand_cleaned # if r_abt.brand_cleaned.startswith(common_str) or r_abt.brand_cleaned.endswith(common_str): # brand_score = 1 # else: # brand_score = 0 # model_score = 0.5 # if r_abt.model_cleaned and r_buy.model_cleaned: # if r_abt.model_cleaned == r_buy.model_cleaned: # model_score = 1 # else: # if len(r_abt.model_cleaned) >= len(r_buy.model_cleaned): # common_str = r_buy.model_cleaned # if r_abt.model_cleaned.startswith(common_str) or r_abt.model_cleaned.endswith(common_str): # model_score = 1 # else: # model_score = 0 # # if brand_score == 1 and model_score == 1: # jaccard_score = 1 # else: # jaccard_score = rltk.jaccard_index_similarity(r_abt.name_tokens, r_buy.name_tokens) # # return [brand_score, model_score, jaccard_score] def generate_feature_vector_raw(r_abt, r_buy): # brand brand_score = None if r_abt.brand_cleaned and r_buy.brand_cleaned: if r_abt.brand_cleaned == r_buy.brand_cleaned: brand_score = 1 # model 1 model_score = None model_marker = 0 if r_abt.model_cleaned and r_buy.model_cleaned: if r_abt.model_cleaned == r_buy.model_cleaned: model_score = 1 else: if len(r_abt.model_cleaned) > len(r_buy.model_cleaned): if r_abt.model_cleaned.startswith(r_buy.model_cleaned) \ or r_abt.model_cleaned.endswith(r_buy.model_cleaned): model_score = 1 else: model_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.model_cleaned) elif len(r_abt.model_cleaned) < len(r_buy.model_cleaned): if r_buy.model_cleaned.startswith(r_abt.model_cleaned) \ or r_buy.model_cleaned.endswith(r_abt.model_cleaned): model_score = 1 else: model_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.model_cleaned) else: model_score = 0 # model 2 model2_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.name_suffix) # name tokens jaccard jaccard_score = rltk.jaccard_index_similarity(r_abt.name_tokens, r_buy.name_tokens) # name tokens tf-idf # t_x = collections.Counter(r_abt.name_tokens) # tf_x = {k: float(v) / len(r_abt.name_tokens) for k, v in t_x.items()} # tfidf_x = {k : tf_x[k] / idf[k] for k, v in tf_x.items()} # t_y = collections.Counter(r_buy.name_tokens) # tf_y = {k: float(v) / len(r_buy.name_tokens) for k, v in t_y.items()} # tfidf_y = {k : tf_y[k] / idf[k] for k, v in tf_y.items()} # tfidf_score = rltk.tf_idf_similarity_by_dict(tfidf_x, tfidf_y) tfidf_score = tfidf.similarity(r_abt.id, r_buy.id) # price if r_abt.price and r_buy.price: price_marker = 1 abt_price = float(r_abt.price) buy_price = float(r_buy.price) if abt_price == 0 and buy_price == 0: price_difference = 0 else: price_difference = float(abs(abt_price - buy_price)) / max(abt_price, buy_price) else: price_marker = 0 price_difference = 0 return [brand_score, model_score, model2_score, jaccard_score, tfidf_score, price_difference, price_marker] def generate_feature_vector(r_abt, r_buy): # brand brand_score = 0.2 brand_marker = 0 if r_abt.brand_cleaned and r_buy.brand_cleaned: if r_abt.brand_cleaned == r_buy.brand_cleaned: brand_score = 1 brand_marker = 1 # model 1 model_score = 0.2 model_marker = 0 if r_abt.model_cleaned and r_buy.model_cleaned: if r_abt.model_cleaned == r_buy.model_cleaned: model_score = 1 model_marker = 1 else: if len(r_abt.model_cleaned) > len(r_buy.model_cleaned): if r_abt.model_cleaned.startswith(r_buy.model_cleaned) \ or r_abt.model_cleaned.endswith(r_buy.model_cleaned): model_score = 1 model_marker = 1 else: model_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.model_cleaned) elif len(r_abt.model_cleaned) < len(r_buy.model_cleaned): if r_buy.model_cleaned.startswith(r_abt.model_cleaned) \ or r_buy.model_cleaned.endswith(r_abt.model_cleaned): model_score = 1 model_marker = 1 else: model_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.model_cleaned) else: model_score = 0 # model 2 model2_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.name_suffix) # name tokens jaccard jaccard_score = rltk.jaccard_index_similarity(r_abt.name_tokens, r_buy.name_tokens) tfidf_score = tfidf.similarity(r_abt.id, r_buy.id) # price if r_abt.price and r_buy.price: price_marker = 1 abt_price = float(r_abt.price) buy_price = float(r_buy.price) if abt_price == 0 and buy_price == 0: price_difference = 0 else: price_difference = float(abs(abt_price - buy_price)) / max(abt_price, buy_price) else: price_marker = 0 price_difference = 0 return [brand_score, brand_marker, model_score, model_marker, model2_score, jaccard_score, tfidf_score, price_difference, price_marker] def non_ml_method(r_abt, r_buy): brand_score = 0 if r_abt.brand_cleaned and r_buy.brand_cleaned: if r_abt.brand_cleaned == r_buy.brand_cleaned: brand_score = 1 model_score = 0 if r_abt.model_cleaned and r_buy.model_cleaned: if r_abt.model_cleaned == r_buy.model_cleaned: model_score = 1 jaccard_score = rltk.jaccard_index_similarity(r_abt.name_tokens, r_buy.name_tokens) if model_score == 1: return True total = brand_score * 0.3 + model_score * 0.3 + jaccard_score * 0.4 return total > 0.45
38.445087
108
0.621711
948
6,651
3.996835
0.074895
0.059119
0.068884
0.12246
0.86408
0.825812
0.811824
0.773291
0.765901
0.73819
0
0.014487
0.294241
6,651
173
109
38.445087
0.792714
0.266576
0
0.833333
1
0
0
0
0
0
0
0
0
1
0.029412
false
0
0.009804
0
0.078431
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
bc37c9139bda28d7dcbe08d115fee4810065a313
14,971
py
Python
environments/CTF/loggingFunctions.py
vanstrn/RL_public
0e971e40e063b17918460e19728f95d7924af8db
[ "MIT" ]
1
2021-03-19T17:57:51.000Z
2021-03-19T17:57:51.000Z
environments/CTF/loggingFunctions.py
vanstrn/RL_public
0e971e40e063b17918460e19728f95d7924af8db
[ "MIT" ]
null
null
null
environments/CTF/loggingFunctions.py
vanstrn/RL_public
0e971e40e063b17918460e19728f95d7924af8db
[ "MIT" ]
null
null
null
import tensorflow as tf import numpy as np import itertools import matplotlib.pyplot as plt import matplotlib.image as mpimg import random class ValueTest_StackedStates(tf.keras.callbacks.Callback): def __init__(self,superEpochs,env,network,imageDir=None,freq=50): self.env = env self.network=network[3] self.imageDir=imageDir self.freq = freq self.superEpochs = superEpochs def on_train_end(self, logs=None): if self.superEpochs%self.freq == 0: s=self.env.reset() rewardMap = np.zeros((s.shape[1],s.shape[2])) for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])): grid = self.env.ConstructSample([i,j]) if grid is None: continue tmp = np.concatenate([grid] * 4, axis=2) [value] = SF4.predict(np.expand_dims(tmp,0)) rewardMap[i,j] = value fig=plt.figure(figsize=(5.5, 8)) fig.add_subplot(2,1,1) plt.title("State") imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1) fig.add_subplot(2,1,2) plt.title("Value Prediction") imgplot = plt.imshow(rewardMap) fig.colorbar(imgplot) plt.savefig(self.imageDir+"/ValuePred"+str(self.superEpochs)+".png") plt.close() class ValueTest(tf.keras.callbacks.Callback): def __init__(self,superEpochs,env,network,imageDir=None,freq=50): self.env = env self.network=network[3] self.imageDir=imageDir self.freq = freq self.superEpochs = superEpochs def on_train_end(self, logs=None): if self.superEpochs%self.freq == 0: s=self.env.reset() rewardMap = np.zeros((s.shape[1],s.shape[2])) for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])): grid = self.env.ConstructSample([i,j]) if grid is None: continue [value] = self.network.predict(np.expand_dims(grid,0)) rewardMap[i,j] = value fig=plt.figure(figsize=(5.5, 8)) fig.add_subplot(2,1,1) plt.title("State") imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1) fig.add_subplot(2,1,2) plt.title("Value Prediction") imgplot = plt.imshow(rewardMap) fig.colorbar(imgplot) plt.savefig(self.imageDir+"/ValuePred"+str(self.superEpochs)+".png") plt.close() class ValueTest_enemies(tf.keras.callbacks.Callback): def __init__(self,superEpochs,env,network,imageDir=None,freq=50): self.env = env self.network=network[3] self.imageDir=imageDir self.freq = freq self.superEpochs = superEpochs def on_train_end(self, logs=None): if self.superEpochs%self.freq == 0: s = self.env.reset() for sample in range(5): enemy_position = [random.randint(0,s.shape[1]-1),random.randint(0,s.shape[2]-1)] while s[0,enemy_position[0],enemy_position[1],3] ==1: enemy_position = [random.randint(0,s.shape[1]-1),random.randint(0,s.shape[2]-1)] rewardMap = np.zeros((s.shape[1],s.shape[2])) for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])): grid = self.env.ConstructSample_e([i,j],enemy_position) if grid is None: continue [value] = self.network.predict(np.expand_dims(grid,0)) rewardMap[i,j] = value fig=plt.figure(figsize=(5.5, 10)) fig.add_subplot(3,1,1) plt.title("State Flags") imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1) fig.add_subplot(3,1,2) plt.title("State Enemies") imgplot = plt.imshow(grid[:,:,4], vmin=-1, vmax=1) fig.add_subplot(3,1,3) plt.title("Value Prediction Epoch "+str(self.superEpochs)) imgplot = plt.imshow(rewardMap) fig.colorbar(imgplot) plt.savefig(self.imageDir+"/ValuePred"+str(self.superEpochs)+"_sample"+str(sample)+".png") plt.close() class ImageGenerator_StackedStates(tf.keras.callbacks.Callback): def __init__(self,env,network,imageDir=None,freq=50): self.env = env self.network=network[0] self.imageDir = imageDir self.freq = freq def on_epoch_end(self,epoch, logs=None): if epoch%self.freq == 0: state = self.env.reset() [state_new,reward] = self.network.predict([np.stack([[0,0,0,0,0]]),state]) fig=plt.figure(figsize=(16, 8)) fig.add_subplot(2,4,1) plt.title("State Territory") imgplot = plt.imshow(state[0,:,:,1], vmin=-2, vmax=2) fig.add_subplot(2,4,2) plt.title("State Flags") imgplot = plt.imshow(state[0,:,:,2], vmin=-10, vmax=10) fig.add_subplot(2,4,3) plt.title("State Obstacles") imgplot = plt.imshow(state[0,:,:,3], vmin=-2, vmax=2) fig.add_subplot(2,4,4) plt.title("State Agents") imgplot = plt.imshow(state[0,:,:,4],vmin=-10, vmax=10) fig.add_subplot(2,4,5) plt.title("Predicted Next State Territory") imgplot = plt.imshow(state_new[0,:,:,1],vmin=-2, vmax=2) fig.add_subplot(2,4,6) plt.title("Predicted Next State Flags") imgplot = plt.imshow(state_new[0,:,:,2],vmin=-10, vmax=10) fig.add_subplot(2,4,7) plt.title("Predicted Next State Obstacles") imgplot = plt.imshow(state_new[0,:,:,3],vmin=-2, vmax=2) fig.add_subplot(2,4,8) plt.title("Predicted Next State Agents") imgplot = plt.imshow(state_new[0,:,:,4],vmin=-10, vmax=10) plt.savefig(self.imageDir+"/StatePredEpoch"+str(epoch)+".png") class ImageGenerator_actions(tf.keras.callbacks.Callback): def __init__(self,env,network,imageDir=None,freq=50): self.env = env self.network=network[0] self.imageDir = imageDir self.freq = freq def on_epoch_end(self,epoch, logs=None): if epoch%self.freq == 0: state = self.env.reset() [state_new,reward] = self.network.predict([np.stack([[0,0,0,0,0]]),state]) fig=plt.figure(figsize=(16, 8)) fig.add_subplot(2,4,1) plt.title("State Territory") imgplot = plt.imshow(state[0,:,:,1], vmin=-2, vmax=2) fig.add_subplot(2,4,2) plt.title("State Flags") imgplot = plt.imshow(state[0,:,:,2], vmin=-10, vmax=10) fig.add_subplot(2,4,3) plt.title("State Obstacles") imgplot = plt.imshow(state[0,:,:,3], vmin=-2, vmax=2) fig.add_subplot(2,4,4) plt.title("State Agents") imgplot = plt.imshow(state[0,:,:,4],vmin=-10, vmax=10) fig.add_subplot(2,4,5) plt.title("Predicted Next State Territory") imgplot = plt.imshow(state_new[0,:,:,1],vmin=-2, vmax=2) fig.add_subplot(2,4,6) plt.title("Predicted Next State Flags") imgplot = plt.imshow(state_new[0,:,:,2],vmin=-10, vmax=10) fig.add_subplot(2,4,7) plt.title("Predicted Next State Obstacles") imgplot = plt.imshow(state_new[0,:,:,3],vmin=-2, vmax=2) fig.add_subplot(2,4,8) plt.title("Predicted Next State Agents") imgplot = plt.imshow(state_new[0,:,:,4],vmin=-10, vmax=10) plt.savefig(self.imageDir+"/StatePredEpoch"+str(epoch)+".png") plt.close() class ImageGenerator(tf.keras.callbacks.Callback): def __init__(self,env,network,imageDir=None,freq=50): self.env = env self.network=network[0] self.imageDir = imageDir self.freq = freq def on_epoch_end(self,epoch, logs=None): if epoch%self.freq == 0: state = self.env.reset() [state_new,reward] = self.network.predict([state]) fig=plt.figure(figsize=(16, 8)) fig.add_subplot(2,4,1) plt.title("State Territory") imgplot = plt.imshow(state[0,:,:,1], vmin=-2, vmax=2) fig.add_subplot(2,4,2) plt.title("State Flags") imgplot = plt.imshow(state[0,:,:,2], vmin=-10, vmax=10) fig.add_subplot(2,4,3) plt.title("State Obstacles") imgplot = plt.imshow(state[0,:,:,3], vmin=-2, vmax=2) fig.add_subplot(2,4,4) plt.title("State Agents") imgplot = plt.imshow(state[0,:,:,4],vmin=-10, vmax=10) fig.add_subplot(2,4,5) plt.title("Predicted Next State Territory") imgplot = plt.imshow(state_new[0,:,:,1],vmin=-2, vmax=2) fig.add_subplot(2,4,6) plt.title("Predicted Next State Flags") imgplot = plt.imshow(state_new[0,:,:,2],vmin=-10, vmax=10) fig.add_subplot(2,4,7) plt.title("Predicted Next State Obstacles") imgplot = plt.imshow(state_new[0,:,:,3],vmin=-2, vmax=2) fig.add_subplot(2,4,8) plt.title("Predicted Next State Agents") imgplot = plt.imshow(state_new[0,:,:,4],vmin=-10, vmax=10) plt.savefig(self.imageDir+"/StatePredEpoch"+str(epoch)+".png") plt.close() class RewardTest(tf.keras.callbacks.Callback): def __init__(self,env,network,imageDir=None,freq=50): self.env = env self.network=network[0] self.imageDir = imageDir self.freq = freq def on_epoch_end(self,epoch, logs=None): if epoch%self.freq == 0: s = self.env.reset() rewardMap = np.zeros((s.shape[1],s.shape[2])) for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])): grid = self.env.ConstructSample([i,j]) if grid is None: continue [state_new,reward] = self.network.predict(np.expand_dims(grid,0)) rewardMap[i,j] = reward fig=plt.figure(figsize=(5.5, 8)) fig.add_subplot(2,1,1) plt.title("State") imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1) fig.add_subplot(2,1,2) plt.title("Reward Prediction Epoch "+str(epoch)) imgplot = plt.imshow(rewardMap) fig.colorbar(imgplot) plt.savefig(self.imageDir+"/RewardPred"+str(epoch)+".png") plt.close() class RewardTest_actions(tf.keras.callbacks.Callback): def __init__(self,env,network,imageDir=None,freq=50): self.env = env self.network=network[0] self.imageDir = imageDir self.freq = freq def on_epoch_end(self,epoch, logs=None): if epoch%self.freq == 0: s = self.env.reset() rewardMap = np.zeros((s.shape[1],s.shape[2])) for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])): grid = self.env.ConstructSample([i,j]) if grid is None: continue [state_new,reward] = self.network.predict([np.stack([[0,0,0,0,0]]),np.expand_dims(grid,0)]) rewardMap[i,j] = reward fig=plt.figure(figsize=(5.5, 8)) fig.add_subplot(2,1,1) plt.title("State") imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1) fig.add_subplot(2,1,2) plt.title("Reward Prediction Epoch "+str(epoch)) imgplot = plt.imshow(rewardMap) fig.colorbar(imgplot) plt.savefig(self.imageDir+"/RewardPred"+str(epoch)+".png") plt.close() class RewardTest_actions_enemies(tf.keras.callbacks.Callback): def __init__(self,env,network,imageDir=None,freq=50): self.env = env self.network=network[0] self.imageDir = imageDir self.freq = freq def on_epoch_end(self,epoch, logs=None): if epoch%self.freq == 0: s = self.env.reset() for sample in range(5): enemy_position = [random.randint(0,s.shape[1]-1),random.randint(0,s.shape[2]-1)] while s[0,enemy_position[0],enemy_position[1],3] ==1: enemy_position = [random.randint(0,s.shape[1]-1),random.randint(0,s.shape[2]-1)] rewardMap = np.zeros((s.shape[1],s.shape[2])) for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])): grid = self.env.ConstructSample_e([i,j],enemy_position) if grid is None: continue [state_new,reward] = self.network.predict([np.stack([[0,0,0,0,0]]),np.expand_dims(grid,0)]) rewardMap[i,j] = reward # fig=plt.figure() fig=plt.figure(figsize=(5.5, 10)) fig.add_subplot(3,1,1) plt.title("State Flags") imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1) fig.add_subplot(3,1,2) plt.title("State Enemies") imgplot = plt.imshow(grid[:,:,4], vmin=-1, vmax=1) fig.add_subplot(3,1,3) plt.title("Reward Prediction Epoch "+str(epoch)) imgplot = plt.imshow(rewardMap) fig.colorbar(imgplot) plt.savefig(self.imageDir+"/RewardPred"+str(epoch)+"_sample"+str(sample)+".png") plt.close() class RewardTest_StackedStates(tf.keras.callbacks.Callback): def __init__(self,env,network,imageDir=None,freq=50): self.env = env self.network=network[0] self.imageDir = imageDir self.freq = freq def on_epoch_end(self,epoch, logs=None): if epoch%self.freq == 0: s = self.env.reset() rewardMap = np.zeros((s.shape[1],s.shape[2])) for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])): grid = self.env.ConstructSample([i,j]) if grid is None: continue tmp = np.concatenate([grid] * 4, axis=2) [state_new,reward] = self.network.predict([np.stack([[1,1,1,1,1]]),np.expand_dims(tmp,0)]) rewardMap[i,j] = reward fig=plt.figure(figsize=(5.5, 8)) fig.add_subplot(2,1,1) plt.title("State") imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1) fig.add_subplot(2,1,2) plt.title("Reward Prediction Epoch "+str(epoch)) imgplot = plt.imshow(rewardMap) fig.colorbar(imgplot) plt.savefig(self.imageDir+"/RewardPred"+str(epoch)+".png") plt.close()
45.643293
111
0.558079
2,022
14,971
4.051929
0.054896
0.057366
0.063469
0.058098
0.974612
0.969974
0.969974
0.961186
0.949469
0.949469
0
0.041284
0.28809
14,971
327
112
45.782875
0.727435
0.001069
0
0.925806
0
0
0.059854
0
0
0
0
0
0
1
0.064516
false
0
0.019355
0
0.116129
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
70c93acd37b7b2c527bb8dbf47471beb4e21276e
68
py
Python
recipes/recipes_emscripten/boost-histogram/test_import_boost_histogram.py
emscripten-forge/recipes
62cb3e146abc8945ac210f38e4e47c080698eae5
[ "MIT" ]
1
2022-03-10T16:50:56.000Z
2022-03-10T16:50:56.000Z
recipes/recipes_emscripten/boost-histogram/test_import_boost_histogram.py
emscripten-forge/recipes
62cb3e146abc8945ac210f38e4e47c080698eae5
[ "MIT" ]
9
2022-03-18T09:26:38.000Z
2022-03-29T09:21:51.000Z
recipes/recipes_emscripten/boost-histogram/test_import_boost_histogram.py
emscripten-forge/recipes
62cb3e146abc8945ac210f38e4e47c080698eae5
[ "MIT" ]
null
null
null
def test_import_boost_histogram(): import boost_histogram
13.6
34
0.75
8
68
5.875
0.625
0.468085
0.851064
0
0
0
0
0
0
0
0
0
0.205882
68
5
35
13.6
0.87037
0
0
0
0
0
0
0
0
0
0
0
0
1
0.5
true
0
1
0
1.5
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
0
1
0
1
0
0
8
cb05f03e29fb61dbf7e515fc6ca4be9e2be0a4f9
2,753
py
Python
gilbo_experiments_trace_diag/griewank-counterex-tensorflow.py
BachiLi/autodiff_comp
d2e33db4bd4ac7630b6ee6cfc56b3411cc37e14d
[ "MIT" ]
2
2020-04-15T21:02:55.000Z
2020-05-07T00:37:21.000Z
gilbo_experiments_trace_diag/griewank-counterex-tensorflow.py
BachiLi/autodiff_comp
d2e33db4bd4ac7630b6ee6cfc56b3411cc37e14d
[ "MIT" ]
null
null
null
gilbo_experiments_trace_diag/griewank-counterex-tensorflow.py
BachiLi/autodiff_comp
d2e33db4bd4ac7630b6ee6cfc56b3411cc37e14d
[ "MIT" ]
1
2020-05-06T17:13:56.000Z
2020-05-06T17:13:56.000Z
import numpy as np import tensorflow as tf import time import math tf.compat.v1.enable_eager_execution() def run_diag_trace(N,K): def f(x): A = tf.linalg.diag(x) v = tf.linalg.trace(A) for i in range(1,K): v += tf.linalg.trace(A) return v def Df(x): grads = None with tf.GradientTape() as tape: tape.watch(x) v = f(x) grads = tape.gradient(v, x) return grads x = tf.random.uniform([N]) f(x) Df(x) start_time = time.perf_counter() v = f(x) stop_time = time.perf_counter() t = stop_time - start_time start_time = time.perf_counter() g = Df(x) stop_time = time.perf_counter() dt = stop_time - start_time return (dt / t), t, dt def xla_diag_trace(N,K): @tf.function( experimental_compile=True, input_signature=[tf.TensorSpec(shape=[N], dtype=tf.float32)] ) def f(x): A = tf.linalg.diag(x) v = tf.linalg.trace(A) for i in range(1,K): v += tf.linalg.trace(A) return v @tf.function( experimental_compile=True, input_signature=[tf.TensorSpec(shape=[N], dtype=tf.float32)] ) def Df(x): grads = None with tf.GradientTape() as tape: tape.watch(x) A = tf.linalg.diag(x) v = tf.linalg.trace(A) for i in range(1,K): v += tf.linalg.trace(A) grads = tape.gradient(v, x) return grads x = tf.random.uniform([N]) f(x) Df(x) start_time = time.perf_counter() v = f(x) stop_time = time.perf_counter() t = stop_time - start_time start_time = time.perf_counter() g = Df(x) stop_time = time.perf_counter() dt = stop_time - start_time return (dt / t), t, dt def xla_diag_dot(N,K): @tf.function( experimental_compile=True, input_signature=[tf.TensorSpec(shape=[N], dtype=tf.float32)] ) def f(x): A = tf.linalg.diag(x) v = tf.tensordot(A[:,0], A[0,:], 1) return v @tf.function( experimental_compile=True, input_signature=[tf.TensorSpec(shape=[N], dtype=tf.float32)] ) def Df(x): grads = None with tf.GradientTape() as tape: tape.watch(x) A = tf.linalg.diag(x) v = tf.tensordot(A[:,0], A[0,:], 1) grads = tape.gradient(v, x) return grads x = tf.random.uniform([N]) f(x) Df(x) start_time = time.perf_counter() v = f(x) stop_time = time.perf_counter() t = stop_time - start_time start_time = time.perf_counter() g = Df(x) stop_time = time.perf_counter() dt = stop_time - start_time return (dt / t), t, dt for K in range(1,11): N = 4000 * K ratio, t, dt = xla_diag_dot(N,K) print(f"Griewank ratio for N = {N}: {ratio}, {t}, {dt}") #print(f" : {t} {dt}") print("\n----\n")
19.664286
64
0.587359
443
2,753
3.528217
0.148984
0.069098
0.092131
0.145873
0.880998
0.871401
0.871401
0.871401
0.871401
0.871401
0
0.012267
0.259717
2,753
139
65
19.805755
0.754661
0.007628
0
0.839623
0
0
0.020191
0
0
0
0
0
0
1
0.084906
false
0
0.037736
0
0.207547
0.018868
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
cb1332c53cf6e76166c0a29efa79493bd05d3f39
158
py
Python
provision/src/util.py
SMISC/logging
b0c98737c46305e3e0718e429635043a4c57e9e0
[ "MIT" ]
null
null
null
provision/src/util.py
SMISC/logging
b0c98737c46305e3e0718e429635043a4c57e9e0
[ "MIT" ]
null
null
null
provision/src/util.py
SMISC/logging
b0c98737c46305e3e0718e429635043a4c57e9e0
[ "MIT" ]
null
null
null
import datetime def twittertime(dt): return (datetime.datetime.strptime(dt, "%a %b %d %H:%M:%S +0000 %Y") - datetime.datetime(1970,1,1)).total_seconds()
31.6
119
0.683544
25
158
4.28
0.76
0.299065
0
0
0
0
0
0
0
0
0
0.071942
0.120253
158
4
120
39.5
0.697842
0
0
0
0
0
0.164557
0
0
0
0
0
0
1
0.333333
false
0
0.333333
0.333333
1
0
1
0
0
null
1
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
1
1
1
0
0
7
cb1a078cdd007b165ff0ae0c8e5db1b2f7e2b2ba
105
py
Python
token_myprint_runtest.py
jamesthiele/acodec
ac93cf5b6f7b1014e89abc47c0857e16d1a1aadf
[ "BSD-2-Clause" ]
null
null
null
token_myprint_runtest.py
jamesthiele/acodec
ac93cf5b6f7b1014e89abc47c0857e16d1a1aadf
[ "BSD-2-Clause" ]
null
null
null
token_myprint_runtest.py
jamesthiele/acodec
ac93cf5b6f7b1014e89abc47c0857e16d1a1aadf
[ "BSD-2-Clause" ]
null
null
null
import os os.system("rm token_myprint_test.pyc") import token_myprint_codec import token_myprint_test
13.125
38
0.838095
17
105
4.823529
0.529412
0.439024
0.390244
0
0
0
0
0
0
0
0
0
0.104762
105
7
39
15
0.87234
0
0
0
0
0
0.240385
0.211538
0
0
0
0
0
1
0
true
0
0.75
0
0.75
0.75
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
0
1
0
7
cb1ab9290d80e85d6655e191541d99167b4c4dff
129,345
py
Python
operators/keycloak-operator/python/pulumi_pulumi_kubernetes_crds_operators_keycloak_operator/keycloak/v1alpha1/outputs.py
pulumi/pulumi-kubernetes-crds
372c4c0182f6b899af82d6edaad521aa14f22150
[ "Apache-2.0" ]
null
null
null
operators/keycloak-operator/python/pulumi_pulumi_kubernetes_crds_operators_keycloak_operator/keycloak/v1alpha1/outputs.py
pulumi/pulumi-kubernetes-crds
372c4c0182f6b899af82d6edaad521aa14f22150
[ "Apache-2.0" ]
2
2020-09-18T17:12:23.000Z
2020-12-30T19:40:56.000Z
operators/keycloak-operator/python/pulumi_pulumi_kubernetes_crds_operators_keycloak_operator/keycloak/v1alpha1/outputs.py
pulumi/pulumi-kubernetes-crds
372c4c0182f6b899af82d6edaad521aa14f22150
[ "Apache-2.0" ]
null
null
null
# coding=utf-8 # *** WARNING: this file was generated by crd2pulumi. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union from ... import _utilities, _tables from . import outputs __all__ = [ 'KeycloakBackupSpec', 'KeycloakBackupSpecAws', 'KeycloakBackupSpecInstanceSelector', 'KeycloakBackupSpecInstanceSelectorMatchExpressions', 'KeycloakBackupStatus', 'KeycloakClientSpec', 'KeycloakClientSpecClient', 'KeycloakClientSpecClientProtocolMappers', 'KeycloakClientSpecRealmSelector', 'KeycloakClientSpecRealmSelectorMatchExpressions', 'KeycloakClientStatus', 'KeycloakRealmSpec', 'KeycloakRealmSpecInstanceSelector', 'KeycloakRealmSpecInstanceSelectorMatchExpressions', 'KeycloakRealmSpecRealm', 'KeycloakRealmSpecRealmClients', 'KeycloakRealmSpecRealmClientsProtocolMappers', 'KeycloakRealmSpecRealmIdentityProviders', 'KeycloakRealmSpecRealmOverrides', 'KeycloakRealmSpecRealmUsers', 'KeycloakRealmSpecRealmUsersCredentials', 'KeycloakRealmSpecRealmUsersFederatedIdentities', 'KeycloakRealmStatus', 'KeycloakSpec', 'KeycloakSpecExternalAccess', 'KeycloakSpecExternalDatabase', 'KeycloakSpecKeycloakDeploymentSpec', 'KeycloakSpecKeycloakDeploymentSpecResources', 'KeycloakSpecMigration', 'KeycloakSpecMigrationBackups', 'KeycloakSpecPodDisruptionBudget', 'KeycloakSpecPostgresDeploymentSpec', 'KeycloakSpecPostgresDeploymentSpecResources', 'KeycloakStatus', 'KeycloakUserSpec', 'KeycloakUserSpecRealmSelector', 'KeycloakUserSpecRealmSelectorMatchExpressions', 'KeycloakUserSpecUser', 'KeycloakUserSpecUserCredentials', 'KeycloakUserSpecUserFederatedIdentities', 'KeycloakUserStatus', ] @pulumi.output_type class KeycloakBackupSpec(dict): """ KeycloakBackupSpec defines the desired state of KeycloakBackup. """ def __init__(__self__, *, aws: Optional['outputs.KeycloakBackupSpecAws'] = None, instance_selector: Optional['outputs.KeycloakBackupSpecInstanceSelector'] = None, restore: Optional[bool] = None, storage_class_name: Optional[str] = None): """ KeycloakBackupSpec defines the desired state of KeycloakBackup. :param 'KeycloakBackupSpecAwsArgs' aws: If provided, an automatic database backup will be created on AWS S3 instead of a local Persistent Volume. If this property is not provided - a local Persistent Volume backup will be chosen. :param 'KeycloakBackupSpecInstanceSelectorArgs' instance_selector: Selector for looking up Keycloak Custom Resources. :param bool restore: Controls automatic restore behavior. Currently not implemented. In the future this will be used to trigger automatic restore for a given KeycloakBackup. Each backup will correspond to a single snapshot of the database (stored either in a Persistent Volume or AWS). If a user wants to restore it, all he/she needs to do is to change this flag to true. Potentially, it will be possible to restore a single backup multiple times. :param str storage_class_name: Name of the StorageClass for Postgresql Backup Persistent Volume Claim """ if aws is not None: pulumi.set(__self__, "aws", aws) if instance_selector is not None: pulumi.set(__self__, "instance_selector", instance_selector) if restore is not None: pulumi.set(__self__, "restore", restore) if storage_class_name is not None: pulumi.set(__self__, "storage_class_name", storage_class_name) @property @pulumi.getter def aws(self) -> Optional['outputs.KeycloakBackupSpecAws']: """ If provided, an automatic database backup will be created on AWS S3 instead of a local Persistent Volume. If this property is not provided - a local Persistent Volume backup will be chosen. """ return pulumi.get(self, "aws") @property @pulumi.getter(name="instanceSelector") def instance_selector(self) -> Optional['outputs.KeycloakBackupSpecInstanceSelector']: """ Selector for looking up Keycloak Custom Resources. """ return pulumi.get(self, "instance_selector") @property @pulumi.getter def restore(self) -> Optional[bool]: """ Controls automatic restore behavior. Currently not implemented. In the future this will be used to trigger automatic restore for a given KeycloakBackup. Each backup will correspond to a single snapshot of the database (stored either in a Persistent Volume or AWS). If a user wants to restore it, all he/she needs to do is to change this flag to true. Potentially, it will be possible to restore a single backup multiple times. """ return pulumi.get(self, "restore") @property @pulumi.getter(name="storageClassName") def storage_class_name(self) -> Optional[str]: """ Name of the StorageClass for Postgresql Backup Persistent Volume Claim """ return pulumi.get(self, "storage_class_name") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakBackupSpecAws(dict): """ If provided, an automatic database backup will be created on AWS S3 instead of a local Persistent Volume. If this property is not provided - a local Persistent Volume backup will be chosen. """ def __init__(__self__, *, credentials_secret_name: Optional[str] = None, encryption_key_secret_name: Optional[str] = None, schedule: Optional[str] = None): """ If provided, an automatic database backup will be created on AWS S3 instead of a local Persistent Volume. If this property is not provided - a local Persistent Volume backup will be chosen. :param str credentials_secret_name: Provides a secret name used for connecting to AWS S3 Service. The secret needs to be in the following form: apiVersion: v1 kind: Secret metadata: name: <Secret name> type: Opaque stringData: AWS_S3_BUCKET_NAME: <S3 Bucket Name> AWS_ACCESS_KEY_ID: <AWS Access Key ID> AWS_SECRET_ACCESS_KEY: <AWS Secret Key> For more information, please refer to the Operator documentation. :param str encryption_key_secret_name: If provided, the database backup will be encrypted. Provides a secret name used for encrypting database data. The secret needs to be in the following form: apiVersion: v1 kind: Secret metadata: name: <Secret name> type: Opaque stringData: GPG_PUBLIC_KEY: <GPG Public Key> GPG_TRUST_MODEL: <GPG Trust Model> GPG_RECIPIENT: <GPG Recipient> For more information, please refer to the Operator documentation. :param str schedule: If specified, it will be used as a schedule for creating a CronJob. """ if credentials_secret_name is not None: pulumi.set(__self__, "credentials_secret_name", credentials_secret_name) if encryption_key_secret_name is not None: pulumi.set(__self__, "encryption_key_secret_name", encryption_key_secret_name) if schedule is not None: pulumi.set(__self__, "schedule", schedule) @property @pulumi.getter(name="credentialsSecretName") def credentials_secret_name(self) -> Optional[str]: """ Provides a secret name used for connecting to AWS S3 Service. The secret needs to be in the following form: apiVersion: v1 kind: Secret metadata: name: <Secret name> type: Opaque stringData: AWS_S3_BUCKET_NAME: <S3 Bucket Name> AWS_ACCESS_KEY_ID: <AWS Access Key ID> AWS_SECRET_ACCESS_KEY: <AWS Secret Key> For more information, please refer to the Operator documentation. """ return pulumi.get(self, "credentials_secret_name") @property @pulumi.getter(name="encryptionKeySecretName") def encryption_key_secret_name(self) -> Optional[str]: """ If provided, the database backup will be encrypted. Provides a secret name used for encrypting database data. The secret needs to be in the following form: apiVersion: v1 kind: Secret metadata: name: <Secret name> type: Opaque stringData: GPG_PUBLIC_KEY: <GPG Public Key> GPG_TRUST_MODEL: <GPG Trust Model> GPG_RECIPIENT: <GPG Recipient> For more information, please refer to the Operator documentation. """ return pulumi.get(self, "encryption_key_secret_name") @property @pulumi.getter def schedule(self) -> Optional[str]: """ If specified, it will be used as a schedule for creating a CronJob. """ return pulumi.get(self, "schedule") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakBackupSpecInstanceSelector(dict): """ Selector for looking up Keycloak Custom Resources. """ def __init__(__self__, *, match_expressions: Optional[Sequence['outputs.KeycloakBackupSpecInstanceSelectorMatchExpressions']] = None, match_labels: Optional[Mapping[str, str]] = None): """ Selector for looking up Keycloak Custom Resources. :param Sequence['KeycloakBackupSpecInstanceSelectorMatchExpressionsArgs'] match_expressions: matchExpressions is a list of label selector requirements. The requirements are ANDed. :param Mapping[str, str] match_labels: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. """ if match_expressions is not None: pulumi.set(__self__, "match_expressions", match_expressions) if match_labels is not None: pulumi.set(__self__, "match_labels", match_labels) @property @pulumi.getter(name="matchExpressions") def match_expressions(self) -> Optional[Sequence['outputs.KeycloakBackupSpecInstanceSelectorMatchExpressions']]: """ matchExpressions is a list of label selector requirements. The requirements are ANDed. """ return pulumi.get(self, "match_expressions") @property @pulumi.getter(name="matchLabels") def match_labels(self) -> Optional[Mapping[str, str]]: """ matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. """ return pulumi.get(self, "match_labels") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakBackupSpecInstanceSelectorMatchExpressions(dict): """ A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. """ def __init__(__self__, *, key: str, operator: str, values: Optional[Sequence[str]] = None): """ A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. :param str key: key is the label key that the selector applies to. :param str operator: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. :param Sequence[str] values: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. """ pulumi.set(__self__, "key", key) pulumi.set(__self__, "operator", operator) if values is not None: pulumi.set(__self__, "values", values) @property @pulumi.getter def key(self) -> str: """ key is the label key that the selector applies to. """ return pulumi.get(self, "key") @property @pulumi.getter def operator(self) -> str: """ operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. """ return pulumi.get(self, "operator") @property @pulumi.getter def values(self) -> Optional[Sequence[str]]: """ values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. """ return pulumi.get(self, "values") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakBackupStatus(dict): """ KeycloakBackupStatus defines the observed state of KeycloakBackup. """ def __init__(__self__, *, message: str, phase: str, ready: bool, secondary_resources: Optional[Mapping[str, Sequence[str]]] = None): """ KeycloakBackupStatus defines the observed state of KeycloakBackup. :param str message: Human-readable message indicating details about current operator phase or error. :param str phase: Current phase of the operator. :param bool ready: True if all resources are in a ready state and all work is done. :param Mapping[str, Sequence[str]] secondary_resources: A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ] """ pulumi.set(__self__, "message", message) pulumi.set(__self__, "phase", phase) pulumi.set(__self__, "ready", ready) if secondary_resources is not None: pulumi.set(__self__, "secondary_resources", secondary_resources) @property @pulumi.getter def message(self) -> str: """ Human-readable message indicating details about current operator phase or error. """ return pulumi.get(self, "message") @property @pulumi.getter def phase(self) -> str: """ Current phase of the operator. """ return pulumi.get(self, "phase") @property @pulumi.getter def ready(self) -> bool: """ True if all resources are in a ready state and all work is done. """ return pulumi.get(self, "ready") @property @pulumi.getter(name="secondaryResources") def secondary_resources(self) -> Optional[Mapping[str, Sequence[str]]]: """ A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ] """ return pulumi.get(self, "secondary_resources") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakClientSpec(dict): """ KeycloakClientSpec defines the desired state of KeycloakClient. """ def __init__(__self__, *, client: 'outputs.KeycloakClientSpecClient', realm_selector: 'outputs.KeycloakClientSpecRealmSelector'): """ KeycloakClientSpec defines the desired state of KeycloakClient. :param 'KeycloakClientSpecClientArgs' client: Keycloak Client REST object. :param 'KeycloakClientSpecRealmSelectorArgs' realm_selector: Selector for looking up KeycloakRealm Custom Resources. """ pulumi.set(__self__, "client", client) pulumi.set(__self__, "realm_selector", realm_selector) @property @pulumi.getter def client(self) -> 'outputs.KeycloakClientSpecClient': """ Keycloak Client REST object. """ return pulumi.get(self, "client") @property @pulumi.getter(name="realmSelector") def realm_selector(self) -> 'outputs.KeycloakClientSpecRealmSelector': """ Selector for looking up KeycloakRealm Custom Resources. """ return pulumi.get(self, "realm_selector") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakClientSpecClient(dict): """ Keycloak Client REST object. """ def __init__(__self__, *, client_id: str, access: Optional[Mapping[str, bool]] = None, admin_url: Optional[str] = None, attributes: Optional[Mapping[str, str]] = None, base_url: Optional[str] = None, bearer_only: Optional[bool] = None, client_authenticator_type: Optional[str] = None, consent_required: Optional[bool] = None, default_roles: Optional[Sequence[str]] = None, description: Optional[str] = None, direct_access_grants_enabled: Optional[bool] = None, enabled: Optional[bool] = None, frontchannel_logout: Optional[bool] = None, full_scope_allowed: Optional[bool] = None, id: Optional[str] = None, implicit_flow_enabled: Optional[bool] = None, name: Optional[str] = None, node_re_registration_timeout: Optional[int] = None, not_before: Optional[int] = None, protocol: Optional[str] = None, protocol_mappers: Optional[Sequence['outputs.KeycloakClientSpecClientProtocolMappers']] = None, public_client: Optional[bool] = None, redirect_uris: Optional[Sequence[str]] = None, root_url: Optional[str] = None, secret: Optional[str] = None, service_accounts_enabled: Optional[bool] = None, standard_flow_enabled: Optional[bool] = None, surrogate_auth_required: Optional[bool] = None, use_template_config: Optional[bool] = None, use_template_mappers: Optional[bool] = None, use_template_scope: Optional[bool] = None, web_origins: Optional[Sequence[str]] = None): """ Keycloak Client REST object. :param str client_id: Client ID. :param Mapping[str, bool] access: Access options. :param str admin_url: Application Admin URL. :param Mapping[str, str] attributes: Client Attributes. :param str base_url: Application base URL. :param bool bearer_only: True if a client supports only Bearer Tokens. :param str client_authenticator_type: What Client authentication type to use. :param bool consent_required: True if Consent Screen is required. :param Sequence[str] default_roles: Default Client roles. :param str description: Client description. :param bool direct_access_grants_enabled: True if Direct Grant is enabled. :param bool enabled: Client enabled flag. :param bool frontchannel_logout: True if this client supports Front Channel logout. :param bool full_scope_allowed: True if Full Scope is allowed. :param str id: Client ID. If not specified, automatically generated. :param bool implicit_flow_enabled: True if Implicit flow is enabled. :param str name: Client name. :param int node_re_registration_timeout: Node registration timeout. :param int not_before: Not Before setting. :param str protocol: Protocol used for this Client. :param Sequence['KeycloakClientSpecClientProtocolMappersArgs'] protocol_mappers: Protocol Mappers. :param bool public_client: True if this is a public Client. :param Sequence[str] redirect_uris: A list of valid Redirection URLs. :param str root_url: Application root URL. :param str secret: Client Secret. The Operator will automatically create a Secret based on this value. :param bool service_accounts_enabled: True if Service Accounts are enabled. :param bool standard_flow_enabled: True if Standard flow is enabled. :param bool surrogate_auth_required: Surrogate Authentication Required option. :param bool use_template_config: True to use a Template Config. :param bool use_template_mappers: True to use Template Mappers. :param bool use_template_scope: True to use Template Scope. :param Sequence[str] web_origins: A list of valid Web Origins. """ pulumi.set(__self__, "client_id", client_id) if access is not None: pulumi.set(__self__, "access", access) if admin_url is not None: pulumi.set(__self__, "admin_url", admin_url) if attributes is not None: pulumi.set(__self__, "attributes", attributes) if base_url is not None: pulumi.set(__self__, "base_url", base_url) if bearer_only is not None: pulumi.set(__self__, "bearer_only", bearer_only) if client_authenticator_type is not None: pulumi.set(__self__, "client_authenticator_type", client_authenticator_type) if consent_required is not None: pulumi.set(__self__, "consent_required", consent_required) if default_roles is not None: pulumi.set(__self__, "default_roles", default_roles) if description is not None: pulumi.set(__self__, "description", description) if direct_access_grants_enabled is not None: pulumi.set(__self__, "direct_access_grants_enabled", direct_access_grants_enabled) if enabled is not None: pulumi.set(__self__, "enabled", enabled) if frontchannel_logout is not None: pulumi.set(__self__, "frontchannel_logout", frontchannel_logout) if full_scope_allowed is not None: pulumi.set(__self__, "full_scope_allowed", full_scope_allowed) if id is not None: pulumi.set(__self__, "id", id) if implicit_flow_enabled is not None: pulumi.set(__self__, "implicit_flow_enabled", implicit_flow_enabled) if name is not None: pulumi.set(__self__, "name", name) if node_re_registration_timeout is not None: pulumi.set(__self__, "node_re_registration_timeout", node_re_registration_timeout) if not_before is not None: pulumi.set(__self__, "not_before", not_before) if protocol is not None: pulumi.set(__self__, "protocol", protocol) if protocol_mappers is not None: pulumi.set(__self__, "protocol_mappers", protocol_mappers) if public_client is not None: pulumi.set(__self__, "public_client", public_client) if redirect_uris is not None: pulumi.set(__self__, "redirect_uris", redirect_uris) if root_url is not None: pulumi.set(__self__, "root_url", root_url) if secret is not None: pulumi.set(__self__, "secret", secret) if service_accounts_enabled is not None: pulumi.set(__self__, "service_accounts_enabled", service_accounts_enabled) if standard_flow_enabled is not None: pulumi.set(__self__, "standard_flow_enabled", standard_flow_enabled) if surrogate_auth_required is not None: pulumi.set(__self__, "surrogate_auth_required", surrogate_auth_required) if use_template_config is not None: pulumi.set(__self__, "use_template_config", use_template_config) if use_template_mappers is not None: pulumi.set(__self__, "use_template_mappers", use_template_mappers) if use_template_scope is not None: pulumi.set(__self__, "use_template_scope", use_template_scope) if web_origins is not None: pulumi.set(__self__, "web_origins", web_origins) @property @pulumi.getter(name="clientId") def client_id(self) -> str: """ Client ID. """ return pulumi.get(self, "client_id") @property @pulumi.getter def access(self) -> Optional[Mapping[str, bool]]: """ Access options. """ return pulumi.get(self, "access") @property @pulumi.getter(name="adminUrl") def admin_url(self) -> Optional[str]: """ Application Admin URL. """ return pulumi.get(self, "admin_url") @property @pulumi.getter def attributes(self) -> Optional[Mapping[str, str]]: """ Client Attributes. """ return pulumi.get(self, "attributes") @property @pulumi.getter(name="baseUrl") def base_url(self) -> Optional[str]: """ Application base URL. """ return pulumi.get(self, "base_url") @property @pulumi.getter(name="bearerOnly") def bearer_only(self) -> Optional[bool]: """ True if a client supports only Bearer Tokens. """ return pulumi.get(self, "bearer_only") @property @pulumi.getter(name="clientAuthenticatorType") def client_authenticator_type(self) -> Optional[str]: """ What Client authentication type to use. """ return pulumi.get(self, "client_authenticator_type") @property @pulumi.getter(name="consentRequired") def consent_required(self) -> Optional[bool]: """ True if Consent Screen is required. """ return pulumi.get(self, "consent_required") @property @pulumi.getter(name="defaultRoles") def default_roles(self) -> Optional[Sequence[str]]: """ Default Client roles. """ return pulumi.get(self, "default_roles") @property @pulumi.getter def description(self) -> Optional[str]: """ Client description. """ return pulumi.get(self, "description") @property @pulumi.getter(name="directAccessGrantsEnabled") def direct_access_grants_enabled(self) -> Optional[bool]: """ True if Direct Grant is enabled. """ return pulumi.get(self, "direct_access_grants_enabled") @property @pulumi.getter def enabled(self) -> Optional[bool]: """ Client enabled flag. """ return pulumi.get(self, "enabled") @property @pulumi.getter(name="frontchannelLogout") def frontchannel_logout(self) -> Optional[bool]: """ True if this client supports Front Channel logout. """ return pulumi.get(self, "frontchannel_logout") @property @pulumi.getter(name="fullScopeAllowed") def full_scope_allowed(self) -> Optional[bool]: """ True if Full Scope is allowed. """ return pulumi.get(self, "full_scope_allowed") @property @pulumi.getter def id(self) -> Optional[str]: """ Client ID. If not specified, automatically generated. """ return pulumi.get(self, "id") @property @pulumi.getter(name="implicitFlowEnabled") def implicit_flow_enabled(self) -> Optional[bool]: """ True if Implicit flow is enabled. """ return pulumi.get(self, "implicit_flow_enabled") @property @pulumi.getter def name(self) -> Optional[str]: """ Client name. """ return pulumi.get(self, "name") @property @pulumi.getter(name="nodeReRegistrationTimeout") def node_re_registration_timeout(self) -> Optional[int]: """ Node registration timeout. """ return pulumi.get(self, "node_re_registration_timeout") @property @pulumi.getter(name="notBefore") def not_before(self) -> Optional[int]: """ Not Before setting. """ return pulumi.get(self, "not_before") @property @pulumi.getter def protocol(self) -> Optional[str]: """ Protocol used for this Client. """ return pulumi.get(self, "protocol") @property @pulumi.getter(name="protocolMappers") def protocol_mappers(self) -> Optional[Sequence['outputs.KeycloakClientSpecClientProtocolMappers']]: """ Protocol Mappers. """ return pulumi.get(self, "protocol_mappers") @property @pulumi.getter(name="publicClient") def public_client(self) -> Optional[bool]: """ True if this is a public Client. """ return pulumi.get(self, "public_client") @property @pulumi.getter(name="redirectUris") def redirect_uris(self) -> Optional[Sequence[str]]: """ A list of valid Redirection URLs. """ return pulumi.get(self, "redirect_uris") @property @pulumi.getter(name="rootUrl") def root_url(self) -> Optional[str]: """ Application root URL. """ return pulumi.get(self, "root_url") @property @pulumi.getter def secret(self) -> Optional[str]: """ Client Secret. The Operator will automatically create a Secret based on this value. """ return pulumi.get(self, "secret") @property @pulumi.getter(name="serviceAccountsEnabled") def service_accounts_enabled(self) -> Optional[bool]: """ True if Service Accounts are enabled. """ return pulumi.get(self, "service_accounts_enabled") @property @pulumi.getter(name="standardFlowEnabled") def standard_flow_enabled(self) -> Optional[bool]: """ True if Standard flow is enabled. """ return pulumi.get(self, "standard_flow_enabled") @property @pulumi.getter(name="surrogateAuthRequired") def surrogate_auth_required(self) -> Optional[bool]: """ Surrogate Authentication Required option. """ return pulumi.get(self, "surrogate_auth_required") @property @pulumi.getter(name="useTemplateConfig") def use_template_config(self) -> Optional[bool]: """ True to use a Template Config. """ return pulumi.get(self, "use_template_config") @property @pulumi.getter(name="useTemplateMappers") def use_template_mappers(self) -> Optional[bool]: """ True to use Template Mappers. """ return pulumi.get(self, "use_template_mappers") @property @pulumi.getter(name="useTemplateScope") def use_template_scope(self) -> Optional[bool]: """ True to use Template Scope. """ return pulumi.get(self, "use_template_scope") @property @pulumi.getter(name="webOrigins") def web_origins(self) -> Optional[Sequence[str]]: """ A list of valid Web Origins. """ return pulumi.get(self, "web_origins") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakClientSpecClientProtocolMappers(dict): def __init__(__self__, *, config: Optional[Mapping[str, str]] = None, consent_required: Optional[bool] = None, consent_text: Optional[str] = None, id: Optional[str] = None, name: Optional[str] = None, protocol: Optional[str] = None, protocol_mapper: Optional[str] = None): """ :param Mapping[str, str] config: Config options. :param bool consent_required: True if Consent Screen is required. :param str consent_text: Text to use for displaying Consent Screen. :param str id: Protocol Mapper ID. :param str name: Protocol Mapper Name. :param str protocol: Protocol to use. :param str protocol_mapper: Protocol Mapper to use """ if config is not None: pulumi.set(__self__, "config", config) if consent_required is not None: pulumi.set(__self__, "consent_required", consent_required) if consent_text is not None: pulumi.set(__self__, "consent_text", consent_text) if id is not None: pulumi.set(__self__, "id", id) if name is not None: pulumi.set(__self__, "name", name) if protocol is not None: pulumi.set(__self__, "protocol", protocol) if protocol_mapper is not None: pulumi.set(__self__, "protocol_mapper", protocol_mapper) @property @pulumi.getter def config(self) -> Optional[Mapping[str, str]]: """ Config options. """ return pulumi.get(self, "config") @property @pulumi.getter(name="consentRequired") def consent_required(self) -> Optional[bool]: """ True if Consent Screen is required. """ return pulumi.get(self, "consent_required") @property @pulumi.getter(name="consentText") def consent_text(self) -> Optional[str]: """ Text to use for displaying Consent Screen. """ return pulumi.get(self, "consent_text") @property @pulumi.getter def id(self) -> Optional[str]: """ Protocol Mapper ID. """ return pulumi.get(self, "id") @property @pulumi.getter def name(self) -> Optional[str]: """ Protocol Mapper Name. """ return pulumi.get(self, "name") @property @pulumi.getter def protocol(self) -> Optional[str]: """ Protocol to use. """ return pulumi.get(self, "protocol") @property @pulumi.getter(name="protocolMapper") def protocol_mapper(self) -> Optional[str]: """ Protocol Mapper to use """ return pulumi.get(self, "protocol_mapper") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakClientSpecRealmSelector(dict): """ Selector for looking up KeycloakRealm Custom Resources. """ def __init__(__self__, *, match_expressions: Optional[Sequence['outputs.KeycloakClientSpecRealmSelectorMatchExpressions']] = None, match_labels: Optional[Mapping[str, str]] = None): """ Selector for looking up KeycloakRealm Custom Resources. :param Sequence['KeycloakClientSpecRealmSelectorMatchExpressionsArgs'] match_expressions: matchExpressions is a list of label selector requirements. The requirements are ANDed. :param Mapping[str, str] match_labels: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. """ if match_expressions is not None: pulumi.set(__self__, "match_expressions", match_expressions) if match_labels is not None: pulumi.set(__self__, "match_labels", match_labels) @property @pulumi.getter(name="matchExpressions") def match_expressions(self) -> Optional[Sequence['outputs.KeycloakClientSpecRealmSelectorMatchExpressions']]: """ matchExpressions is a list of label selector requirements. The requirements are ANDed. """ return pulumi.get(self, "match_expressions") @property @pulumi.getter(name="matchLabels") def match_labels(self) -> Optional[Mapping[str, str]]: """ matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. """ return pulumi.get(self, "match_labels") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakClientSpecRealmSelectorMatchExpressions(dict): """ A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. """ def __init__(__self__, *, key: str, operator: str, values: Optional[Sequence[str]] = None): """ A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. :param str key: key is the label key that the selector applies to. :param str operator: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. :param Sequence[str] values: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. """ pulumi.set(__self__, "key", key) pulumi.set(__self__, "operator", operator) if values is not None: pulumi.set(__self__, "values", values) @property @pulumi.getter def key(self) -> str: """ key is the label key that the selector applies to. """ return pulumi.get(self, "key") @property @pulumi.getter def operator(self) -> str: """ operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. """ return pulumi.get(self, "operator") @property @pulumi.getter def values(self) -> Optional[Sequence[str]]: """ values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. """ return pulumi.get(self, "values") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakClientStatus(dict): """ KeycloakClientStatus defines the observed state of KeycloakClient """ def __init__(__self__, *, message: str, phase: str, ready: bool, secondary_resources: Optional[Mapping[str, Sequence[str]]] = None): """ KeycloakClientStatus defines the observed state of KeycloakClient :param str message: Human-readable message indicating details about current operator phase or error. :param str phase: Current phase of the operator. :param bool ready: True if all resources are in a ready state and all work is done. :param Mapping[str, Sequence[str]] secondary_resources: A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ] """ pulumi.set(__self__, "message", message) pulumi.set(__self__, "phase", phase) pulumi.set(__self__, "ready", ready) if secondary_resources is not None: pulumi.set(__self__, "secondary_resources", secondary_resources) @property @pulumi.getter def message(self) -> str: """ Human-readable message indicating details about current operator phase or error. """ return pulumi.get(self, "message") @property @pulumi.getter def phase(self) -> str: """ Current phase of the operator. """ return pulumi.get(self, "phase") @property @pulumi.getter def ready(self) -> bool: """ True if all resources are in a ready state and all work is done. """ return pulumi.get(self, "ready") @property @pulumi.getter(name="secondaryResources") def secondary_resources(self) -> Optional[Mapping[str, Sequence[str]]]: """ A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ] """ return pulumi.get(self, "secondary_resources") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpec(dict): """ KeycloakRealmSpec defines the desired state of KeycloakRealm. """ def __init__(__self__, *, realm: 'outputs.KeycloakRealmSpecRealm', instance_selector: Optional['outputs.KeycloakRealmSpecInstanceSelector'] = None, realm_overrides: Optional[Sequence['outputs.KeycloakRealmSpecRealmOverrides']] = None): """ KeycloakRealmSpec defines the desired state of KeycloakRealm. :param 'KeycloakRealmSpecRealmArgs' realm: Keycloak Realm REST object. :param 'KeycloakRealmSpecInstanceSelectorArgs' instance_selector: Selector for looking up Keycloak Custom Resources. :param Sequence['KeycloakRealmSpecRealmOverridesArgs'] realm_overrides: A list of overrides to the default Realm behavior. """ pulumi.set(__self__, "realm", realm) if instance_selector is not None: pulumi.set(__self__, "instance_selector", instance_selector) if realm_overrides is not None: pulumi.set(__self__, "realm_overrides", realm_overrides) @property @pulumi.getter def realm(self) -> 'outputs.KeycloakRealmSpecRealm': """ Keycloak Realm REST object. """ return pulumi.get(self, "realm") @property @pulumi.getter(name="instanceSelector") def instance_selector(self) -> Optional['outputs.KeycloakRealmSpecInstanceSelector']: """ Selector for looking up Keycloak Custom Resources. """ return pulumi.get(self, "instance_selector") @property @pulumi.getter(name="realmOverrides") def realm_overrides(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmOverrides']]: """ A list of overrides to the default Realm behavior. """ return pulumi.get(self, "realm_overrides") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecInstanceSelector(dict): """ Selector for looking up Keycloak Custom Resources. """ def __init__(__self__, *, match_expressions: Optional[Sequence['outputs.KeycloakRealmSpecInstanceSelectorMatchExpressions']] = None, match_labels: Optional[Mapping[str, str]] = None): """ Selector for looking up Keycloak Custom Resources. :param Sequence['KeycloakRealmSpecInstanceSelectorMatchExpressionsArgs'] match_expressions: matchExpressions is a list of label selector requirements. The requirements are ANDed. :param Mapping[str, str] match_labels: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. """ if match_expressions is not None: pulumi.set(__self__, "match_expressions", match_expressions) if match_labels is not None: pulumi.set(__self__, "match_labels", match_labels) @property @pulumi.getter(name="matchExpressions") def match_expressions(self) -> Optional[Sequence['outputs.KeycloakRealmSpecInstanceSelectorMatchExpressions']]: """ matchExpressions is a list of label selector requirements. The requirements are ANDed. """ return pulumi.get(self, "match_expressions") @property @pulumi.getter(name="matchLabels") def match_labels(self) -> Optional[Mapping[str, str]]: """ matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. """ return pulumi.get(self, "match_labels") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecInstanceSelectorMatchExpressions(dict): """ A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. """ def __init__(__self__, *, key: str, operator: str, values: Optional[Sequence[str]] = None): """ A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. :param str key: key is the label key that the selector applies to. :param str operator: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. :param Sequence[str] values: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. """ pulumi.set(__self__, "key", key) pulumi.set(__self__, "operator", operator) if values is not None: pulumi.set(__self__, "values", values) @property @pulumi.getter def key(self) -> str: """ key is the label key that the selector applies to. """ return pulumi.get(self, "key") @property @pulumi.getter def operator(self) -> str: """ operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. """ return pulumi.get(self, "operator") @property @pulumi.getter def values(self) -> Optional[Sequence[str]]: """ values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. """ return pulumi.get(self, "values") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecRealm(dict): """ Keycloak Realm REST object. """ def __init__(__self__, *, realm: str, admin_events_details_enabled: Optional[bool] = None, admin_events_enabled: Optional[bool] = None, clients: Optional[Sequence['outputs.KeycloakRealmSpecRealmClients']] = None, display_name: Optional[str] = None, enabled: Optional[bool] = None, events_enabled: Optional[bool] = None, events_listeners: Optional[Sequence[str]] = None, id: Optional[str] = None, identity_providers: Optional[Sequence['outputs.KeycloakRealmSpecRealmIdentityProviders']] = None, users: Optional[Sequence['outputs.KeycloakRealmSpecRealmUsers']] = None): """ Keycloak Realm REST object. :param str realm: Realm name. :param bool admin_events_details_enabled: Enable admin events details TODO: change to values and use kubebuilder default annotation once supported :param bool admin_events_enabled: Enable events recording TODO: change to values and use kubebuilder default annotation once supported :param Sequence['KeycloakRealmSpecRealmClientsArgs'] clients: A set of Keycloak Clients. :param str display_name: Realm display name. :param bool enabled: Realm enabled flag. :param bool events_enabled: Enable events recording TODO: change to values and use kubebuilder default annotation once supported :param Sequence[str] events_listeners: A set of Event Listeners. :param Sequence['KeycloakRealmSpecRealmIdentityProvidersArgs'] identity_providers: A set of Identity Providers. :param Sequence['KeycloakRealmSpecRealmUsersArgs'] users: A set of Keycloak Users. """ pulumi.set(__self__, "realm", realm) if admin_events_details_enabled is not None: pulumi.set(__self__, "admin_events_details_enabled", admin_events_details_enabled) if admin_events_enabled is not None: pulumi.set(__self__, "admin_events_enabled", admin_events_enabled) if clients is not None: pulumi.set(__self__, "clients", clients) if display_name is not None: pulumi.set(__self__, "display_name", display_name) if enabled is not None: pulumi.set(__self__, "enabled", enabled) if events_enabled is not None: pulumi.set(__self__, "events_enabled", events_enabled) if events_listeners is not None: pulumi.set(__self__, "events_listeners", events_listeners) if id is not None: pulumi.set(__self__, "id", id) if identity_providers is not None: pulumi.set(__self__, "identity_providers", identity_providers) if users is not None: pulumi.set(__self__, "users", users) @property @pulumi.getter def realm(self) -> str: """ Realm name. """ return pulumi.get(self, "realm") @property @pulumi.getter(name="adminEventsDetailsEnabled") def admin_events_details_enabled(self) -> Optional[bool]: """ Enable admin events details TODO: change to values and use kubebuilder default annotation once supported """ return pulumi.get(self, "admin_events_details_enabled") @property @pulumi.getter(name="adminEventsEnabled") def admin_events_enabled(self) -> Optional[bool]: """ Enable events recording TODO: change to values and use kubebuilder default annotation once supported """ return pulumi.get(self, "admin_events_enabled") @property @pulumi.getter def clients(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmClients']]: """ A set of Keycloak Clients. """ return pulumi.get(self, "clients") @property @pulumi.getter(name="displayName") def display_name(self) -> Optional[str]: """ Realm display name. """ return pulumi.get(self, "display_name") @property @pulumi.getter def enabled(self) -> Optional[bool]: """ Realm enabled flag. """ return pulumi.get(self, "enabled") @property @pulumi.getter(name="eventsEnabled") def events_enabled(self) -> Optional[bool]: """ Enable events recording TODO: change to values and use kubebuilder default annotation once supported """ return pulumi.get(self, "events_enabled") @property @pulumi.getter(name="eventsListeners") def events_listeners(self) -> Optional[Sequence[str]]: """ A set of Event Listeners. """ return pulumi.get(self, "events_listeners") @property @pulumi.getter def id(self) -> Optional[str]: return pulumi.get(self, "id") @property @pulumi.getter(name="identityProviders") def identity_providers(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmIdentityProviders']]: """ A set of Identity Providers. """ return pulumi.get(self, "identity_providers") @property @pulumi.getter def users(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmUsers']]: """ A set of Keycloak Users. """ return pulumi.get(self, "users") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecRealmClients(dict): def __init__(__self__, *, client_id: str, access: Optional[Mapping[str, bool]] = None, admin_url: Optional[str] = None, attributes: Optional[Mapping[str, str]] = None, base_url: Optional[str] = None, bearer_only: Optional[bool] = None, client_authenticator_type: Optional[str] = None, consent_required: Optional[bool] = None, default_roles: Optional[Sequence[str]] = None, description: Optional[str] = None, direct_access_grants_enabled: Optional[bool] = None, enabled: Optional[bool] = None, frontchannel_logout: Optional[bool] = None, full_scope_allowed: Optional[bool] = None, id: Optional[str] = None, implicit_flow_enabled: Optional[bool] = None, name: Optional[str] = None, node_re_registration_timeout: Optional[int] = None, not_before: Optional[int] = None, protocol: Optional[str] = None, protocol_mappers: Optional[Sequence['outputs.KeycloakRealmSpecRealmClientsProtocolMappers']] = None, public_client: Optional[bool] = None, redirect_uris: Optional[Sequence[str]] = None, root_url: Optional[str] = None, secret: Optional[str] = None, service_accounts_enabled: Optional[bool] = None, standard_flow_enabled: Optional[bool] = None, surrogate_auth_required: Optional[bool] = None, use_template_config: Optional[bool] = None, use_template_mappers: Optional[bool] = None, use_template_scope: Optional[bool] = None, web_origins: Optional[Sequence[str]] = None): """ :param str client_id: Client ID. :param Mapping[str, bool] access: Access options. :param str admin_url: Application Admin URL. :param Mapping[str, str] attributes: Client Attributes. :param str base_url: Application base URL. :param bool bearer_only: True if a client supports only Bearer Tokens. :param str client_authenticator_type: What Client authentication type to use. :param bool consent_required: True if Consent Screen is required. :param Sequence[str] default_roles: Default Client roles. :param str description: Client description. :param bool direct_access_grants_enabled: True if Direct Grant is enabled. :param bool enabled: Client enabled flag. :param bool frontchannel_logout: True if this client supports Front Channel logout. :param bool full_scope_allowed: True if Full Scope is allowed. :param str id: Client ID. If not specified, automatically generated. :param bool implicit_flow_enabled: True if Implicit flow is enabled. :param str name: Client name. :param int node_re_registration_timeout: Node registration timeout. :param int not_before: Not Before setting. :param str protocol: Protocol used for this Client. :param Sequence['KeycloakRealmSpecRealmClientsProtocolMappersArgs'] protocol_mappers: Protocol Mappers. :param bool public_client: True if this is a public Client. :param Sequence[str] redirect_uris: A list of valid Redirection URLs. :param str root_url: Application root URL. :param str secret: Client Secret. The Operator will automatically create a Secret based on this value. :param bool service_accounts_enabled: True if Service Accounts are enabled. :param bool standard_flow_enabled: True if Standard flow is enabled. :param bool surrogate_auth_required: Surrogate Authentication Required option. :param bool use_template_config: True to use a Template Config. :param bool use_template_mappers: True to use Template Mappers. :param bool use_template_scope: True to use Template Scope. :param Sequence[str] web_origins: A list of valid Web Origins. """ pulumi.set(__self__, "client_id", client_id) if access is not None: pulumi.set(__self__, "access", access) if admin_url is not None: pulumi.set(__self__, "admin_url", admin_url) if attributes is not None: pulumi.set(__self__, "attributes", attributes) if base_url is not None: pulumi.set(__self__, "base_url", base_url) if bearer_only is not None: pulumi.set(__self__, "bearer_only", bearer_only) if client_authenticator_type is not None: pulumi.set(__self__, "client_authenticator_type", client_authenticator_type) if consent_required is not None: pulumi.set(__self__, "consent_required", consent_required) if default_roles is not None: pulumi.set(__self__, "default_roles", default_roles) if description is not None: pulumi.set(__self__, "description", description) if direct_access_grants_enabled is not None: pulumi.set(__self__, "direct_access_grants_enabled", direct_access_grants_enabled) if enabled is not None: pulumi.set(__self__, "enabled", enabled) if frontchannel_logout is not None: pulumi.set(__self__, "frontchannel_logout", frontchannel_logout) if full_scope_allowed is not None: pulumi.set(__self__, "full_scope_allowed", full_scope_allowed) if id is not None: pulumi.set(__self__, "id", id) if implicit_flow_enabled is not None: pulumi.set(__self__, "implicit_flow_enabled", implicit_flow_enabled) if name is not None: pulumi.set(__self__, "name", name) if node_re_registration_timeout is not None: pulumi.set(__self__, "node_re_registration_timeout", node_re_registration_timeout) if not_before is not None: pulumi.set(__self__, "not_before", not_before) if protocol is not None: pulumi.set(__self__, "protocol", protocol) if protocol_mappers is not None: pulumi.set(__self__, "protocol_mappers", protocol_mappers) if public_client is not None: pulumi.set(__self__, "public_client", public_client) if redirect_uris is not None: pulumi.set(__self__, "redirect_uris", redirect_uris) if root_url is not None: pulumi.set(__self__, "root_url", root_url) if secret is not None: pulumi.set(__self__, "secret", secret) if service_accounts_enabled is not None: pulumi.set(__self__, "service_accounts_enabled", service_accounts_enabled) if standard_flow_enabled is not None: pulumi.set(__self__, "standard_flow_enabled", standard_flow_enabled) if surrogate_auth_required is not None: pulumi.set(__self__, "surrogate_auth_required", surrogate_auth_required) if use_template_config is not None: pulumi.set(__self__, "use_template_config", use_template_config) if use_template_mappers is not None: pulumi.set(__self__, "use_template_mappers", use_template_mappers) if use_template_scope is not None: pulumi.set(__self__, "use_template_scope", use_template_scope) if web_origins is not None: pulumi.set(__self__, "web_origins", web_origins) @property @pulumi.getter(name="clientId") def client_id(self) -> str: """ Client ID. """ return pulumi.get(self, "client_id") @property @pulumi.getter def access(self) -> Optional[Mapping[str, bool]]: """ Access options. """ return pulumi.get(self, "access") @property @pulumi.getter(name="adminUrl") def admin_url(self) -> Optional[str]: """ Application Admin URL. """ return pulumi.get(self, "admin_url") @property @pulumi.getter def attributes(self) -> Optional[Mapping[str, str]]: """ Client Attributes. """ return pulumi.get(self, "attributes") @property @pulumi.getter(name="baseUrl") def base_url(self) -> Optional[str]: """ Application base URL. """ return pulumi.get(self, "base_url") @property @pulumi.getter(name="bearerOnly") def bearer_only(self) -> Optional[bool]: """ True if a client supports only Bearer Tokens. """ return pulumi.get(self, "bearer_only") @property @pulumi.getter(name="clientAuthenticatorType") def client_authenticator_type(self) -> Optional[str]: """ What Client authentication type to use. """ return pulumi.get(self, "client_authenticator_type") @property @pulumi.getter(name="consentRequired") def consent_required(self) -> Optional[bool]: """ True if Consent Screen is required. """ return pulumi.get(self, "consent_required") @property @pulumi.getter(name="defaultRoles") def default_roles(self) -> Optional[Sequence[str]]: """ Default Client roles. """ return pulumi.get(self, "default_roles") @property @pulumi.getter def description(self) -> Optional[str]: """ Client description. """ return pulumi.get(self, "description") @property @pulumi.getter(name="directAccessGrantsEnabled") def direct_access_grants_enabled(self) -> Optional[bool]: """ True if Direct Grant is enabled. """ return pulumi.get(self, "direct_access_grants_enabled") @property @pulumi.getter def enabled(self) -> Optional[bool]: """ Client enabled flag. """ return pulumi.get(self, "enabled") @property @pulumi.getter(name="frontchannelLogout") def frontchannel_logout(self) -> Optional[bool]: """ True if this client supports Front Channel logout. """ return pulumi.get(self, "frontchannel_logout") @property @pulumi.getter(name="fullScopeAllowed") def full_scope_allowed(self) -> Optional[bool]: """ True if Full Scope is allowed. """ return pulumi.get(self, "full_scope_allowed") @property @pulumi.getter def id(self) -> Optional[str]: """ Client ID. If not specified, automatically generated. """ return pulumi.get(self, "id") @property @pulumi.getter(name="implicitFlowEnabled") def implicit_flow_enabled(self) -> Optional[bool]: """ True if Implicit flow is enabled. """ return pulumi.get(self, "implicit_flow_enabled") @property @pulumi.getter def name(self) -> Optional[str]: """ Client name. """ return pulumi.get(self, "name") @property @pulumi.getter(name="nodeReRegistrationTimeout") def node_re_registration_timeout(self) -> Optional[int]: """ Node registration timeout. """ return pulumi.get(self, "node_re_registration_timeout") @property @pulumi.getter(name="notBefore") def not_before(self) -> Optional[int]: """ Not Before setting. """ return pulumi.get(self, "not_before") @property @pulumi.getter def protocol(self) -> Optional[str]: """ Protocol used for this Client. """ return pulumi.get(self, "protocol") @property @pulumi.getter(name="protocolMappers") def protocol_mappers(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmClientsProtocolMappers']]: """ Protocol Mappers. """ return pulumi.get(self, "protocol_mappers") @property @pulumi.getter(name="publicClient") def public_client(self) -> Optional[bool]: """ True if this is a public Client. """ return pulumi.get(self, "public_client") @property @pulumi.getter(name="redirectUris") def redirect_uris(self) -> Optional[Sequence[str]]: """ A list of valid Redirection URLs. """ return pulumi.get(self, "redirect_uris") @property @pulumi.getter(name="rootUrl") def root_url(self) -> Optional[str]: """ Application root URL. """ return pulumi.get(self, "root_url") @property @pulumi.getter def secret(self) -> Optional[str]: """ Client Secret. The Operator will automatically create a Secret based on this value. """ return pulumi.get(self, "secret") @property @pulumi.getter(name="serviceAccountsEnabled") def service_accounts_enabled(self) -> Optional[bool]: """ True if Service Accounts are enabled. """ return pulumi.get(self, "service_accounts_enabled") @property @pulumi.getter(name="standardFlowEnabled") def standard_flow_enabled(self) -> Optional[bool]: """ True if Standard flow is enabled. """ return pulumi.get(self, "standard_flow_enabled") @property @pulumi.getter(name="surrogateAuthRequired") def surrogate_auth_required(self) -> Optional[bool]: """ Surrogate Authentication Required option. """ return pulumi.get(self, "surrogate_auth_required") @property @pulumi.getter(name="useTemplateConfig") def use_template_config(self) -> Optional[bool]: """ True to use a Template Config. """ return pulumi.get(self, "use_template_config") @property @pulumi.getter(name="useTemplateMappers") def use_template_mappers(self) -> Optional[bool]: """ True to use Template Mappers. """ return pulumi.get(self, "use_template_mappers") @property @pulumi.getter(name="useTemplateScope") def use_template_scope(self) -> Optional[bool]: """ True to use Template Scope. """ return pulumi.get(self, "use_template_scope") @property @pulumi.getter(name="webOrigins") def web_origins(self) -> Optional[Sequence[str]]: """ A list of valid Web Origins. """ return pulumi.get(self, "web_origins") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecRealmClientsProtocolMappers(dict): def __init__(__self__, *, config: Optional[Mapping[str, str]] = None, consent_required: Optional[bool] = None, consent_text: Optional[str] = None, id: Optional[str] = None, name: Optional[str] = None, protocol: Optional[str] = None, protocol_mapper: Optional[str] = None): """ :param Mapping[str, str] config: Config options. :param bool consent_required: True if Consent Screen is required. :param str consent_text: Text to use for displaying Consent Screen. :param str id: Protocol Mapper ID. :param str name: Protocol Mapper Name. :param str protocol: Protocol to use. :param str protocol_mapper: Protocol Mapper to use """ if config is not None: pulumi.set(__self__, "config", config) if consent_required is not None: pulumi.set(__self__, "consent_required", consent_required) if consent_text is not None: pulumi.set(__self__, "consent_text", consent_text) if id is not None: pulumi.set(__self__, "id", id) if name is not None: pulumi.set(__self__, "name", name) if protocol is not None: pulumi.set(__self__, "protocol", protocol) if protocol_mapper is not None: pulumi.set(__self__, "protocol_mapper", protocol_mapper) @property @pulumi.getter def config(self) -> Optional[Mapping[str, str]]: """ Config options. """ return pulumi.get(self, "config") @property @pulumi.getter(name="consentRequired") def consent_required(self) -> Optional[bool]: """ True if Consent Screen is required. """ return pulumi.get(self, "consent_required") @property @pulumi.getter(name="consentText") def consent_text(self) -> Optional[str]: """ Text to use for displaying Consent Screen. """ return pulumi.get(self, "consent_text") @property @pulumi.getter def id(self) -> Optional[str]: """ Protocol Mapper ID. """ return pulumi.get(self, "id") @property @pulumi.getter def name(self) -> Optional[str]: """ Protocol Mapper Name. """ return pulumi.get(self, "name") @property @pulumi.getter def protocol(self) -> Optional[str]: """ Protocol to use. """ return pulumi.get(self, "protocol") @property @pulumi.getter(name="protocolMapper") def protocol_mapper(self) -> Optional[str]: """ Protocol Mapper to use """ return pulumi.get(self, "protocol_mapper") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecRealmIdentityProviders(dict): def __init__(__self__, *, add_read_token_role_on_create: Optional[bool] = None, alias: Optional[str] = None, config: Optional[Mapping[str, str]] = None, display_name: Optional[str] = None, enabled: Optional[bool] = None, first_broker_login_flow_alias: Optional[str] = None, internal_id: Optional[str] = None, link_only: Optional[bool] = None, post_broker_login_flow_alias: Optional[str] = None, provider_id: Optional[str] = None, store_token: Optional[bool] = None, trust_email: Optional[bool] = None): """ :param bool add_read_token_role_on_create: Adds Read Token role when creating this Identity Provider. :param str alias: Identity Provider Alias. :param Mapping[str, str] config: Identity Provider config. :param str display_name: Identity Provider Display Name. :param bool enabled: Identity Provider enabled flag. :param str first_broker_login_flow_alias: Identity Provider First Broker Login Flow Alias. :param str internal_id: Identity Provider Internal ID. :param bool link_only: Identity Provider Link Only setting. :param str post_broker_login_flow_alias: Identity Provider Post Broker Login Flow Alias. :param str provider_id: Identity Provider ID. :param bool store_token: Identity Provider Store to Token. :param bool trust_email: Identity Provider Trust Email. """ if add_read_token_role_on_create is not None: pulumi.set(__self__, "add_read_token_role_on_create", add_read_token_role_on_create) if alias is not None: pulumi.set(__self__, "alias", alias) if config is not None: pulumi.set(__self__, "config", config) if display_name is not None: pulumi.set(__self__, "display_name", display_name) if enabled is not None: pulumi.set(__self__, "enabled", enabled) if first_broker_login_flow_alias is not None: pulumi.set(__self__, "first_broker_login_flow_alias", first_broker_login_flow_alias) if internal_id is not None: pulumi.set(__self__, "internal_id", internal_id) if link_only is not None: pulumi.set(__self__, "link_only", link_only) if post_broker_login_flow_alias is not None: pulumi.set(__self__, "post_broker_login_flow_alias", post_broker_login_flow_alias) if provider_id is not None: pulumi.set(__self__, "provider_id", provider_id) if store_token is not None: pulumi.set(__self__, "store_token", store_token) if trust_email is not None: pulumi.set(__self__, "trust_email", trust_email) @property @pulumi.getter(name="addReadTokenRoleOnCreate") def add_read_token_role_on_create(self) -> Optional[bool]: """ Adds Read Token role when creating this Identity Provider. """ return pulumi.get(self, "add_read_token_role_on_create") @property @pulumi.getter def alias(self) -> Optional[str]: """ Identity Provider Alias. """ return pulumi.get(self, "alias") @property @pulumi.getter def config(self) -> Optional[Mapping[str, str]]: """ Identity Provider config. """ return pulumi.get(self, "config") @property @pulumi.getter(name="displayName") def display_name(self) -> Optional[str]: """ Identity Provider Display Name. """ return pulumi.get(self, "display_name") @property @pulumi.getter def enabled(self) -> Optional[bool]: """ Identity Provider enabled flag. """ return pulumi.get(self, "enabled") @property @pulumi.getter(name="firstBrokerLoginFlowAlias") def first_broker_login_flow_alias(self) -> Optional[str]: """ Identity Provider First Broker Login Flow Alias. """ return pulumi.get(self, "first_broker_login_flow_alias") @property @pulumi.getter(name="internalId") def internal_id(self) -> Optional[str]: """ Identity Provider Internal ID. """ return pulumi.get(self, "internal_id") @property @pulumi.getter(name="linkOnly") def link_only(self) -> Optional[bool]: """ Identity Provider Link Only setting. """ return pulumi.get(self, "link_only") @property @pulumi.getter(name="postBrokerLoginFlowAlias") def post_broker_login_flow_alias(self) -> Optional[str]: """ Identity Provider Post Broker Login Flow Alias. """ return pulumi.get(self, "post_broker_login_flow_alias") @property @pulumi.getter(name="providerId") def provider_id(self) -> Optional[str]: """ Identity Provider ID. """ return pulumi.get(self, "provider_id") @property @pulumi.getter(name="storeToken") def store_token(self) -> Optional[bool]: """ Identity Provider Store to Token. """ return pulumi.get(self, "store_token") @property @pulumi.getter(name="trustEmail") def trust_email(self) -> Optional[bool]: """ Identity Provider Trust Email. """ return pulumi.get(self, "trust_email") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecRealmOverrides(dict): def __init__(__self__, *, for_flow: Optional[str] = None, identity_provider: Optional[str] = None): """ :param str for_flow: Flow to be overridden. :param str identity_provider: Identity Provider to be overridden. """ if for_flow is not None: pulumi.set(__self__, "for_flow", for_flow) if identity_provider is not None: pulumi.set(__self__, "identity_provider", identity_provider) @property @pulumi.getter(name="forFlow") def for_flow(self) -> Optional[str]: """ Flow to be overridden. """ return pulumi.get(self, "for_flow") @property @pulumi.getter(name="identityProvider") def identity_provider(self) -> Optional[str]: """ Identity Provider to be overridden. """ return pulumi.get(self, "identity_provider") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecRealmUsers(dict): def __init__(__self__, *, attributes: Optional[Mapping[str, Sequence[str]]] = None, client_roles: Optional[Mapping[str, Sequence[str]]] = None, credentials: Optional[Sequence['outputs.KeycloakRealmSpecRealmUsersCredentials']] = None, email: Optional[str] = None, email_verified: Optional[bool] = None, enabled: Optional[bool] = None, federated_identities: Optional[Sequence['outputs.KeycloakRealmSpecRealmUsersFederatedIdentities']] = None, first_name: Optional[str] = None, groups: Optional[Sequence[str]] = None, id: Optional[str] = None, last_name: Optional[str] = None, realm_roles: Optional[Sequence[str]] = None, required_actions: Optional[Sequence[str]] = None, username: Optional[str] = None): """ :param Mapping[str, Sequence[str]] attributes: A set of Attributes. :param Mapping[str, Sequence[str]] client_roles: A set of Client Roles. :param Sequence['KeycloakRealmSpecRealmUsersCredentialsArgs'] credentials: A set of Credentials. :param str email: Email. :param bool email_verified: True if email has already been verified. :param bool enabled: User enabled flag. :param Sequence['KeycloakRealmSpecRealmUsersFederatedIdentitiesArgs'] federated_identities: A set of Federated Identities. :param str first_name: First Name. :param Sequence[str] groups: A set of Groups. :param str id: User ID. :param str last_name: Last Name. :param Sequence[str] realm_roles: A set of Realm Roles. :param Sequence[str] required_actions: A set of Required Actions. :param str username: User Name. """ if attributes is not None: pulumi.set(__self__, "attributes", attributes) if client_roles is not None: pulumi.set(__self__, "client_roles", client_roles) if credentials is not None: pulumi.set(__self__, "credentials", credentials) if email is not None: pulumi.set(__self__, "email", email) if email_verified is not None: pulumi.set(__self__, "email_verified", email_verified) if enabled is not None: pulumi.set(__self__, "enabled", enabled) if federated_identities is not None: pulumi.set(__self__, "federated_identities", federated_identities) if first_name is not None: pulumi.set(__self__, "first_name", first_name) if groups is not None: pulumi.set(__self__, "groups", groups) if id is not None: pulumi.set(__self__, "id", id) if last_name is not None: pulumi.set(__self__, "last_name", last_name) if realm_roles is not None: pulumi.set(__self__, "realm_roles", realm_roles) if required_actions is not None: pulumi.set(__self__, "required_actions", required_actions) if username is not None: pulumi.set(__self__, "username", username) @property @pulumi.getter def attributes(self) -> Optional[Mapping[str, Sequence[str]]]: """ A set of Attributes. """ return pulumi.get(self, "attributes") @property @pulumi.getter(name="clientRoles") def client_roles(self) -> Optional[Mapping[str, Sequence[str]]]: """ A set of Client Roles. """ return pulumi.get(self, "client_roles") @property @pulumi.getter def credentials(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmUsersCredentials']]: """ A set of Credentials. """ return pulumi.get(self, "credentials") @property @pulumi.getter def email(self) -> Optional[str]: """ Email. """ return pulumi.get(self, "email") @property @pulumi.getter(name="emailVerified") def email_verified(self) -> Optional[bool]: """ True if email has already been verified. """ return pulumi.get(self, "email_verified") @property @pulumi.getter def enabled(self) -> Optional[bool]: """ User enabled flag. """ return pulumi.get(self, "enabled") @property @pulumi.getter(name="federatedIdentities") def federated_identities(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmUsersFederatedIdentities']]: """ A set of Federated Identities. """ return pulumi.get(self, "federated_identities") @property @pulumi.getter(name="firstName") def first_name(self) -> Optional[str]: """ First Name. """ return pulumi.get(self, "first_name") @property @pulumi.getter def groups(self) -> Optional[Sequence[str]]: """ A set of Groups. """ return pulumi.get(self, "groups") @property @pulumi.getter def id(self) -> Optional[str]: """ User ID. """ return pulumi.get(self, "id") @property @pulumi.getter(name="lastName") def last_name(self) -> Optional[str]: """ Last Name. """ return pulumi.get(self, "last_name") @property @pulumi.getter(name="realmRoles") def realm_roles(self) -> Optional[Sequence[str]]: """ A set of Realm Roles. """ return pulumi.get(self, "realm_roles") @property @pulumi.getter(name="requiredActions") def required_actions(self) -> Optional[Sequence[str]]: """ A set of Required Actions. """ return pulumi.get(self, "required_actions") @property @pulumi.getter def username(self) -> Optional[str]: """ User Name. """ return pulumi.get(self, "username") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecRealmUsersCredentials(dict): def __init__(__self__, *, temporary: Optional[bool] = None, type: Optional[str] = None, value: Optional[str] = None): """ :param bool temporary: True if this credential object is temporary. :param str type: Credential Type. :param str value: Credential Value. """ if temporary is not None: pulumi.set(__self__, "temporary", temporary) if type is not None: pulumi.set(__self__, "type", type) if value is not None: pulumi.set(__self__, "value", value) @property @pulumi.getter def temporary(self) -> Optional[bool]: """ True if this credential object is temporary. """ return pulumi.get(self, "temporary") @property @pulumi.getter def type(self) -> Optional[str]: """ Credential Type. """ return pulumi.get(self, "type") @property @pulumi.getter def value(self) -> Optional[str]: """ Credential Value. """ return pulumi.get(self, "value") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmSpecRealmUsersFederatedIdentities(dict): def __init__(__self__, *, identity_provider: Optional[str] = None, user_id: Optional[str] = None, user_name: Optional[str] = None): """ :param str identity_provider: Federated Identity Provider. :param str user_id: Federated Identity User ID. :param str user_name: Federated Identity User Name. """ if identity_provider is not None: pulumi.set(__self__, "identity_provider", identity_provider) if user_id is not None: pulumi.set(__self__, "user_id", user_id) if user_name is not None: pulumi.set(__self__, "user_name", user_name) @property @pulumi.getter(name="identityProvider") def identity_provider(self) -> Optional[str]: """ Federated Identity Provider. """ return pulumi.get(self, "identity_provider") @property @pulumi.getter(name="userId") def user_id(self) -> Optional[str]: """ Federated Identity User ID. """ return pulumi.get(self, "user_id") @property @pulumi.getter(name="userName") def user_name(self) -> Optional[str]: """ Federated Identity User Name. """ return pulumi.get(self, "user_name") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakRealmStatus(dict): """ KeycloakRealmStatus defines the observed state of KeycloakRealm """ def __init__(__self__, *, login_url: str, message: str, phase: str, ready: bool, secondary_resources: Optional[Mapping[str, Sequence[str]]] = None): """ KeycloakRealmStatus defines the observed state of KeycloakRealm :param str login_url: TODO :param str message: Human-readable message indicating details about current operator phase or error. :param str phase: Current phase of the operator. :param bool ready: True if all resources are in a ready state and all work is done. :param Mapping[str, Sequence[str]] secondary_resources: A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ] """ pulumi.set(__self__, "login_url", login_url) pulumi.set(__self__, "message", message) pulumi.set(__self__, "phase", phase) pulumi.set(__self__, "ready", ready) if secondary_resources is not None: pulumi.set(__self__, "secondary_resources", secondary_resources) @property @pulumi.getter(name="loginURL") def login_url(self) -> str: """ TODO """ return pulumi.get(self, "login_url") @property @pulumi.getter def message(self) -> str: """ Human-readable message indicating details about current operator phase or error. """ return pulumi.get(self, "message") @property @pulumi.getter def phase(self) -> str: """ Current phase of the operator. """ return pulumi.get(self, "phase") @property @pulumi.getter def ready(self) -> bool: """ True if all resources are in a ready state and all work is done. """ return pulumi.get(self, "ready") @property @pulumi.getter(name="secondaryResources") def secondary_resources(self) -> Optional[Mapping[str, Sequence[str]]]: """ A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ] """ return pulumi.get(self, "secondary_resources") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpec(dict): """ KeycloakSpec defines the desired state of Keycloak. """ def __init__(__self__, *, extensions: Optional[Sequence[str]] = None, external_access: Optional['outputs.KeycloakSpecExternalAccess'] = None, external_database: Optional['outputs.KeycloakSpecExternalDatabase'] = None, instances: Optional[int] = None, keycloak_deployment_spec: Optional['outputs.KeycloakSpecKeycloakDeploymentSpec'] = None, migration: Optional['outputs.KeycloakSpecMigration'] = None, pod_disruption_budget: Optional['outputs.KeycloakSpecPodDisruptionBudget'] = None, postgres_deployment_spec: Optional['outputs.KeycloakSpecPostgresDeploymentSpec'] = None, profile: Optional[str] = None, storage_class_name: Optional[str] = None): """ KeycloakSpec defines the desired state of Keycloak. :param Sequence[str] extensions: A list of extensions, where each one is a URL to a JAR files that will be deployed in Keycloak. :param 'KeycloakSpecExternalAccessArgs' external_access: Controls external Ingress/Route settings. :param 'KeycloakSpecExternalDatabaseArgs' external_database: Controls external database settings. Using an external database requires providing a secret containing credentials as well as connection details. Here's an example of such secret: apiVersion: v1 kind: Secret metadata: name: keycloak-db-secret namespace: keycloak stringData: POSTGRES_DATABASE: <Database Name> POSTGRES_EXTERNAL_ADDRESS: <External Database IP or URL (resolvable by K8s)> POSTGRES_EXTERNAL_PORT: <External Database Port> # Strongly recommended to use <'Keycloak CR Name'-postgresql> POSTGRES_HOST: <Database Service Name> POSTGRES_PASSWORD: <Database Password> # Required for AWS Backup functionality POSTGRES_SUPERUSER: true POSTGRES_USERNAME: <Database Username> type: Opaque Both POSTGRES_EXTERNAL_ADDRESS and POSTGRES_EXTERNAL_PORT are specifically required for creating connection to the external database. The secret name is created using the following convention: <Custom Resource Name>-db-secret For more information, please refer to the Operator documentation. :param int instances: Number of Keycloak instances in HA mode. Default is 1. :param 'KeycloakSpecKeycloakDeploymentSpecArgs' keycloak_deployment_spec: Resources (Requests and Limits) for KeycloakDeployment. :param 'KeycloakSpecMigrationArgs' migration: Specify Migration configuration :param 'KeycloakSpecPodDisruptionBudgetArgs' pod_disruption_budget: Specify PodDisruptionBudget configuration. :param 'KeycloakSpecPostgresDeploymentSpecArgs' postgres_deployment_spec: Resources (Requests and Limits) for PostgresDeployment. :param str profile: Profile used for controlling Operator behavior. Default is empty. :param str storage_class_name: Name of the StorageClass for Postgresql Persistent Volume Claim """ if extensions is not None: pulumi.set(__self__, "extensions", extensions) if external_access is not None: pulumi.set(__self__, "external_access", external_access) if external_database is not None: pulumi.set(__self__, "external_database", external_database) if instances is not None: pulumi.set(__self__, "instances", instances) if keycloak_deployment_spec is not None: pulumi.set(__self__, "keycloak_deployment_spec", keycloak_deployment_spec) if migration is not None: pulumi.set(__self__, "migration", migration) if pod_disruption_budget is not None: pulumi.set(__self__, "pod_disruption_budget", pod_disruption_budget) if postgres_deployment_spec is not None: pulumi.set(__self__, "postgres_deployment_spec", postgres_deployment_spec) if profile is not None: pulumi.set(__self__, "profile", profile) if storage_class_name is not None: pulumi.set(__self__, "storage_class_name", storage_class_name) @property @pulumi.getter def extensions(self) -> Optional[Sequence[str]]: """ A list of extensions, where each one is a URL to a JAR files that will be deployed in Keycloak. """ return pulumi.get(self, "extensions") @property @pulumi.getter(name="externalAccess") def external_access(self) -> Optional['outputs.KeycloakSpecExternalAccess']: """ Controls external Ingress/Route settings. """ return pulumi.get(self, "external_access") @property @pulumi.getter(name="externalDatabase") def external_database(self) -> Optional['outputs.KeycloakSpecExternalDatabase']: """ Controls external database settings. Using an external database requires providing a secret containing credentials as well as connection details. Here's an example of such secret: apiVersion: v1 kind: Secret metadata: name: keycloak-db-secret namespace: keycloak stringData: POSTGRES_DATABASE: <Database Name> POSTGRES_EXTERNAL_ADDRESS: <External Database IP or URL (resolvable by K8s)> POSTGRES_EXTERNAL_PORT: <External Database Port> # Strongly recommended to use <'Keycloak CR Name'-postgresql> POSTGRES_HOST: <Database Service Name> POSTGRES_PASSWORD: <Database Password> # Required for AWS Backup functionality POSTGRES_SUPERUSER: true POSTGRES_USERNAME: <Database Username> type: Opaque Both POSTGRES_EXTERNAL_ADDRESS and POSTGRES_EXTERNAL_PORT are specifically required for creating connection to the external database. The secret name is created using the following convention: <Custom Resource Name>-db-secret For more information, please refer to the Operator documentation. """ return pulumi.get(self, "external_database") @property @pulumi.getter def instances(self) -> Optional[int]: """ Number of Keycloak instances in HA mode. Default is 1. """ return pulumi.get(self, "instances") @property @pulumi.getter(name="keycloakDeploymentSpec") def keycloak_deployment_spec(self) -> Optional['outputs.KeycloakSpecKeycloakDeploymentSpec']: """ Resources (Requests and Limits) for KeycloakDeployment. """ return pulumi.get(self, "keycloak_deployment_spec") @property @pulumi.getter def migration(self) -> Optional['outputs.KeycloakSpecMigration']: """ Specify Migration configuration """ return pulumi.get(self, "migration") @property @pulumi.getter(name="podDisruptionBudget") def pod_disruption_budget(self) -> Optional['outputs.KeycloakSpecPodDisruptionBudget']: """ Specify PodDisruptionBudget configuration. """ return pulumi.get(self, "pod_disruption_budget") @property @pulumi.getter(name="postgresDeploymentSpec") def postgres_deployment_spec(self) -> Optional['outputs.KeycloakSpecPostgresDeploymentSpec']: """ Resources (Requests and Limits) for PostgresDeployment. """ return pulumi.get(self, "postgres_deployment_spec") @property @pulumi.getter def profile(self) -> Optional[str]: """ Profile used for controlling Operator behavior. Default is empty. """ return pulumi.get(self, "profile") @property @pulumi.getter(name="storageClassName") def storage_class_name(self) -> Optional[str]: """ Name of the StorageClass for Postgresql Persistent Volume Claim """ return pulumi.get(self, "storage_class_name") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecExternalAccess(dict): """ Controls external Ingress/Route settings. """ def __init__(__self__, *, enabled: Optional[bool] = None, tls_termination: Optional[str] = None): """ Controls external Ingress/Route settings. :param bool enabled: If set to true, the Operator will create an Ingress or a Route pointing to Keycloak. :param str tls_termination: TLS Termination type for the external access. Setting this field to "reencrypt" will terminate TLS on the Ingress/Route level. Setting this field to "passthrough" will send encrypted traffic to the Pod. If unspecified, defaults to "reencrypt". Note, that this setting has no effect on Ingress as Ingress TLS settings are not reconciled by this operator. In other words, Ingress TLS configuration is the same in both cases and it is up to the user to configure TLS section of the Ingress. """ if enabled is not None: pulumi.set(__self__, "enabled", enabled) if tls_termination is not None: pulumi.set(__self__, "tls_termination", tls_termination) @property @pulumi.getter def enabled(self) -> Optional[bool]: """ If set to true, the Operator will create an Ingress or a Route pointing to Keycloak. """ return pulumi.get(self, "enabled") @property @pulumi.getter(name="tlsTermination") def tls_termination(self) -> Optional[str]: """ TLS Termination type for the external access. Setting this field to "reencrypt" will terminate TLS on the Ingress/Route level. Setting this field to "passthrough" will send encrypted traffic to the Pod. If unspecified, defaults to "reencrypt". Note, that this setting has no effect on Ingress as Ingress TLS settings are not reconciled by this operator. In other words, Ingress TLS configuration is the same in both cases and it is up to the user to configure TLS section of the Ingress. """ return pulumi.get(self, "tls_termination") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecExternalDatabase(dict): """ Controls external database settings. Using an external database requires providing a secret containing credentials as well as connection details. Here's an example of such secret: apiVersion: v1 kind: Secret metadata: name: keycloak-db-secret namespace: keycloak stringData: POSTGRES_DATABASE: <Database Name> POSTGRES_EXTERNAL_ADDRESS: <External Database IP or URL (resolvable by K8s)> POSTGRES_EXTERNAL_PORT: <External Database Port> # Strongly recommended to use <'Keycloak CR Name'-postgresql> POSTGRES_HOST: <Database Service Name> POSTGRES_PASSWORD: <Database Password> # Required for AWS Backup functionality POSTGRES_SUPERUSER: true POSTGRES_USERNAME: <Database Username> type: Opaque Both POSTGRES_EXTERNAL_ADDRESS and POSTGRES_EXTERNAL_PORT are specifically required for creating connection to the external database. The secret name is created using the following convention: <Custom Resource Name>-db-secret For more information, please refer to the Operator documentation. """ def __init__(__self__, *, enabled: Optional[bool] = None): """ Controls external database settings. Using an external database requires providing a secret containing credentials as well as connection details. Here's an example of such secret: apiVersion: v1 kind: Secret metadata: name: keycloak-db-secret namespace: keycloak stringData: POSTGRES_DATABASE: <Database Name> POSTGRES_EXTERNAL_ADDRESS: <External Database IP or URL (resolvable by K8s)> POSTGRES_EXTERNAL_PORT: <External Database Port> # Strongly recommended to use <'Keycloak CR Name'-postgresql> POSTGRES_HOST: <Database Service Name> POSTGRES_PASSWORD: <Database Password> # Required for AWS Backup functionality POSTGRES_SUPERUSER: true POSTGRES_USERNAME: <Database Username> type: Opaque Both POSTGRES_EXTERNAL_ADDRESS and POSTGRES_EXTERNAL_PORT are specifically required for creating connection to the external database. The secret name is created using the following convention: <Custom Resource Name>-db-secret For more information, please refer to the Operator documentation. :param bool enabled: If set to true, the Operator will use an external database. pointing to Keycloak. """ if enabled is not None: pulumi.set(__self__, "enabled", enabled) @property @pulumi.getter def enabled(self) -> Optional[bool]: """ If set to true, the Operator will use an external database. pointing to Keycloak. """ return pulumi.get(self, "enabled") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecKeycloakDeploymentSpec(dict): """ Resources (Requests and Limits) for KeycloakDeployment. """ def __init__(__self__, *, resources: Optional['outputs.KeycloakSpecKeycloakDeploymentSpecResources'] = None): """ Resources (Requests and Limits) for KeycloakDeployment. :param 'KeycloakSpecKeycloakDeploymentSpecResourcesArgs' resources: Resources (Requests and Limits) for the Pods. """ if resources is not None: pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> Optional['outputs.KeycloakSpecKeycloakDeploymentSpecResources']: """ Resources (Requests and Limits) for the Pods. """ return pulumi.get(self, "resources") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecKeycloakDeploymentSpecResources(dict): """ Resources (Requests and Limits) for the Pods. """ def __init__(__self__, *, limits: Optional[Mapping[str, str]] = None, requests: Optional[Mapping[str, str]] = None): """ Resources (Requests and Limits) for the Pods. :param Mapping[str, str] limits: Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ :param Mapping[str, str] requests: Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ """ if limits is not None: pulumi.set(__self__, "limits", limits) if requests is not None: pulumi.set(__self__, "requests", requests) @property @pulumi.getter def limits(self) -> Optional[Mapping[str, str]]: """ Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ """ return pulumi.get(self, "limits") @property @pulumi.getter def requests(self) -> Optional[Mapping[str, str]]: """ Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ """ return pulumi.get(self, "requests") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecMigration(dict): """ Specify Migration configuration """ def __init__(__self__, *, backups: Optional['outputs.KeycloakSpecMigrationBackups'] = None): """ Specify Migration configuration :param 'KeycloakSpecMigrationBackupsArgs' backups: Set it to config backup policy for migration """ if backups is not None: pulumi.set(__self__, "backups", backups) @property @pulumi.getter def backups(self) -> Optional['outputs.KeycloakSpecMigrationBackups']: """ Set it to config backup policy for migration """ return pulumi.get(self, "backups") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecMigrationBackups(dict): """ Set it to config backup policy for migration """ def __init__(__self__, *, enabled: Optional[bool] = None): """ Set it to config backup policy for migration :param bool enabled: If set to true, the operator will do database backup before doing migration """ if enabled is not None: pulumi.set(__self__, "enabled", enabled) @property @pulumi.getter def enabled(self) -> Optional[bool]: """ If set to true, the operator will do database backup before doing migration """ return pulumi.get(self, "enabled") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecPodDisruptionBudget(dict): """ Specify PodDisruptionBudget configuration. """ def __init__(__self__, *, enabled: Optional[bool] = None): """ Specify PodDisruptionBudget configuration. :param bool enabled: If set to true, the operator will create a PodDistruptionBudget for the Keycloak deployment and set its `maxUnavailable` value to 1. """ if enabled is not None: pulumi.set(__self__, "enabled", enabled) @property @pulumi.getter def enabled(self) -> Optional[bool]: """ If set to true, the operator will create a PodDistruptionBudget for the Keycloak deployment and set its `maxUnavailable` value to 1. """ return pulumi.get(self, "enabled") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecPostgresDeploymentSpec(dict): """ Resources (Requests and Limits) for PostgresDeployment. """ def __init__(__self__, *, resources: Optional['outputs.KeycloakSpecPostgresDeploymentSpecResources'] = None): """ Resources (Requests and Limits) for PostgresDeployment. :param 'KeycloakSpecPostgresDeploymentSpecResourcesArgs' resources: Resources (Requests and Limits) for the Pods. """ if resources is not None: pulumi.set(__self__, "resources", resources) @property @pulumi.getter def resources(self) -> Optional['outputs.KeycloakSpecPostgresDeploymentSpecResources']: """ Resources (Requests and Limits) for the Pods. """ return pulumi.get(self, "resources") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakSpecPostgresDeploymentSpecResources(dict): """ Resources (Requests and Limits) for the Pods. """ def __init__(__self__, *, limits: Optional[Mapping[str, str]] = None, requests: Optional[Mapping[str, str]] = None): """ Resources (Requests and Limits) for the Pods. :param Mapping[str, str] limits: Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ :param Mapping[str, str] requests: Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ """ if limits is not None: pulumi.set(__self__, "limits", limits) if requests is not None: pulumi.set(__self__, "requests", requests) @property @pulumi.getter def limits(self) -> Optional[Mapping[str, str]]: """ Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ """ return pulumi.get(self, "limits") @property @pulumi.getter def requests(self) -> Optional[Mapping[str, str]]: """ Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ """ return pulumi.get(self, "requests") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakStatus(dict): """ KeycloakStatus defines the observed state of Keycloak. """ def __init__(__self__, *, credential_secret: str, internal_url: str, message: str, phase: str, ready: bool, version: str, secondary_resources: Optional[Mapping[str, Sequence[str]]] = None): """ KeycloakStatus defines the observed state of Keycloak. :param str credential_secret: The secret where the admin credentials are to be found. :param str internal_url: Service IP and Port for in-cluster access to the keycloak instance. :param str message: Human-readable message indicating details about current operator phase or error. :param str phase: Current phase of the operator. :param bool ready: True if all resources are in a ready state and all work is done. :param str version: Version of Keycloak or RHSSO running on the cluster. :param Mapping[str, Sequence[str]] secondary_resources: A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ]. """ pulumi.set(__self__, "credential_secret", credential_secret) pulumi.set(__self__, "internal_url", internal_url) pulumi.set(__self__, "message", message) pulumi.set(__self__, "phase", phase) pulumi.set(__self__, "ready", ready) pulumi.set(__self__, "version", version) if secondary_resources is not None: pulumi.set(__self__, "secondary_resources", secondary_resources) @property @pulumi.getter(name="credentialSecret") def credential_secret(self) -> str: """ The secret where the admin credentials are to be found. """ return pulumi.get(self, "credential_secret") @property @pulumi.getter(name="internalURL") def internal_url(self) -> str: """ Service IP and Port for in-cluster access to the keycloak instance. """ return pulumi.get(self, "internal_url") @property @pulumi.getter def message(self) -> str: """ Human-readable message indicating details about current operator phase or error. """ return pulumi.get(self, "message") @property @pulumi.getter def phase(self) -> str: """ Current phase of the operator. """ return pulumi.get(self, "phase") @property @pulumi.getter def ready(self) -> bool: """ True if all resources are in a ready state and all work is done. """ return pulumi.get(self, "ready") @property @pulumi.getter def version(self) -> str: """ Version of Keycloak or RHSSO running on the cluster. """ return pulumi.get(self, "version") @property @pulumi.getter(name="secondaryResources") def secondary_resources(self) -> Optional[Mapping[str, Sequence[str]]]: """ A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ]. """ return pulumi.get(self, "secondary_resources") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakUserSpec(dict): """ KeycloakUserSpec defines the desired state of KeycloakUser. """ def __init__(__self__, *, user: 'outputs.KeycloakUserSpecUser', realm_selector: Optional['outputs.KeycloakUserSpecRealmSelector'] = None): """ KeycloakUserSpec defines the desired state of KeycloakUser. :param 'KeycloakUserSpecUserArgs' user: Keycloak User REST object. :param 'KeycloakUserSpecRealmSelectorArgs' realm_selector: Selector for looking up KeycloakRealm Custom Resources. """ pulumi.set(__self__, "user", user) if realm_selector is not None: pulumi.set(__self__, "realm_selector", realm_selector) @property @pulumi.getter def user(self) -> 'outputs.KeycloakUserSpecUser': """ Keycloak User REST object. """ return pulumi.get(self, "user") @property @pulumi.getter(name="realmSelector") def realm_selector(self) -> Optional['outputs.KeycloakUserSpecRealmSelector']: """ Selector for looking up KeycloakRealm Custom Resources. """ return pulumi.get(self, "realm_selector") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakUserSpecRealmSelector(dict): """ Selector for looking up KeycloakRealm Custom Resources. """ def __init__(__self__, *, match_expressions: Optional[Sequence['outputs.KeycloakUserSpecRealmSelectorMatchExpressions']] = None, match_labels: Optional[Mapping[str, str]] = None): """ Selector for looking up KeycloakRealm Custom Resources. :param Sequence['KeycloakUserSpecRealmSelectorMatchExpressionsArgs'] match_expressions: matchExpressions is a list of label selector requirements. The requirements are ANDed. :param Mapping[str, str] match_labels: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. """ if match_expressions is not None: pulumi.set(__self__, "match_expressions", match_expressions) if match_labels is not None: pulumi.set(__self__, "match_labels", match_labels) @property @pulumi.getter(name="matchExpressions") def match_expressions(self) -> Optional[Sequence['outputs.KeycloakUserSpecRealmSelectorMatchExpressions']]: """ matchExpressions is a list of label selector requirements. The requirements are ANDed. """ return pulumi.get(self, "match_expressions") @property @pulumi.getter(name="matchLabels") def match_labels(self) -> Optional[Mapping[str, str]]: """ matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. """ return pulumi.get(self, "match_labels") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakUserSpecRealmSelectorMatchExpressions(dict): """ A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. """ def __init__(__self__, *, key: str, operator: str, values: Optional[Sequence[str]] = None): """ A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. :param str key: key is the label key that the selector applies to. :param str operator: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. :param Sequence[str] values: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. """ pulumi.set(__self__, "key", key) pulumi.set(__self__, "operator", operator) if values is not None: pulumi.set(__self__, "values", values) @property @pulumi.getter def key(self) -> str: """ key is the label key that the selector applies to. """ return pulumi.get(self, "key") @property @pulumi.getter def operator(self) -> str: """ operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. """ return pulumi.get(self, "operator") @property @pulumi.getter def values(self) -> Optional[Sequence[str]]: """ values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. """ return pulumi.get(self, "values") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakUserSpecUser(dict): """ Keycloak User REST object. """ def __init__(__self__, *, attributes: Optional[Mapping[str, Sequence[str]]] = None, client_roles: Optional[Mapping[str, Sequence[str]]] = None, credentials: Optional[Sequence['outputs.KeycloakUserSpecUserCredentials']] = None, email: Optional[str] = None, email_verified: Optional[bool] = None, enabled: Optional[bool] = None, federated_identities: Optional[Sequence['outputs.KeycloakUserSpecUserFederatedIdentities']] = None, first_name: Optional[str] = None, groups: Optional[Sequence[str]] = None, id: Optional[str] = None, last_name: Optional[str] = None, realm_roles: Optional[Sequence[str]] = None, required_actions: Optional[Sequence[str]] = None, username: Optional[str] = None): """ Keycloak User REST object. :param Mapping[str, Sequence[str]] attributes: A set of Attributes. :param Mapping[str, Sequence[str]] client_roles: A set of Client Roles. :param Sequence['KeycloakUserSpecUserCredentialsArgs'] credentials: A set of Credentials. :param str email: Email. :param bool email_verified: True if email has already been verified. :param bool enabled: User enabled flag. :param Sequence['KeycloakUserSpecUserFederatedIdentitiesArgs'] federated_identities: A set of Federated Identities. :param str first_name: First Name. :param Sequence[str] groups: A set of Groups. :param str id: User ID. :param str last_name: Last Name. :param Sequence[str] realm_roles: A set of Realm Roles. :param Sequence[str] required_actions: A set of Required Actions. :param str username: User Name. """ if attributes is not None: pulumi.set(__self__, "attributes", attributes) if client_roles is not None: pulumi.set(__self__, "client_roles", client_roles) if credentials is not None: pulumi.set(__self__, "credentials", credentials) if email is not None: pulumi.set(__self__, "email", email) if email_verified is not None: pulumi.set(__self__, "email_verified", email_verified) if enabled is not None: pulumi.set(__self__, "enabled", enabled) if federated_identities is not None: pulumi.set(__self__, "federated_identities", federated_identities) if first_name is not None: pulumi.set(__self__, "first_name", first_name) if groups is not None: pulumi.set(__self__, "groups", groups) if id is not None: pulumi.set(__self__, "id", id) if last_name is not None: pulumi.set(__self__, "last_name", last_name) if realm_roles is not None: pulumi.set(__self__, "realm_roles", realm_roles) if required_actions is not None: pulumi.set(__self__, "required_actions", required_actions) if username is not None: pulumi.set(__self__, "username", username) @property @pulumi.getter def attributes(self) -> Optional[Mapping[str, Sequence[str]]]: """ A set of Attributes. """ return pulumi.get(self, "attributes") @property @pulumi.getter(name="clientRoles") def client_roles(self) -> Optional[Mapping[str, Sequence[str]]]: """ A set of Client Roles. """ return pulumi.get(self, "client_roles") @property @pulumi.getter def credentials(self) -> Optional[Sequence['outputs.KeycloakUserSpecUserCredentials']]: """ A set of Credentials. """ return pulumi.get(self, "credentials") @property @pulumi.getter def email(self) -> Optional[str]: """ Email. """ return pulumi.get(self, "email") @property @pulumi.getter(name="emailVerified") def email_verified(self) -> Optional[bool]: """ True if email has already been verified. """ return pulumi.get(self, "email_verified") @property @pulumi.getter def enabled(self) -> Optional[bool]: """ User enabled flag. """ return pulumi.get(self, "enabled") @property @pulumi.getter(name="federatedIdentities") def federated_identities(self) -> Optional[Sequence['outputs.KeycloakUserSpecUserFederatedIdentities']]: """ A set of Federated Identities. """ return pulumi.get(self, "federated_identities") @property @pulumi.getter(name="firstName") def first_name(self) -> Optional[str]: """ First Name. """ return pulumi.get(self, "first_name") @property @pulumi.getter def groups(self) -> Optional[Sequence[str]]: """ A set of Groups. """ return pulumi.get(self, "groups") @property @pulumi.getter def id(self) -> Optional[str]: """ User ID. """ return pulumi.get(self, "id") @property @pulumi.getter(name="lastName") def last_name(self) -> Optional[str]: """ Last Name. """ return pulumi.get(self, "last_name") @property @pulumi.getter(name="realmRoles") def realm_roles(self) -> Optional[Sequence[str]]: """ A set of Realm Roles. """ return pulumi.get(self, "realm_roles") @property @pulumi.getter(name="requiredActions") def required_actions(self) -> Optional[Sequence[str]]: """ A set of Required Actions. """ return pulumi.get(self, "required_actions") @property @pulumi.getter def username(self) -> Optional[str]: """ User Name. """ return pulumi.get(self, "username") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakUserSpecUserCredentials(dict): def __init__(__self__, *, temporary: Optional[bool] = None, type: Optional[str] = None, value: Optional[str] = None): """ :param bool temporary: True if this credential object is temporary. :param str type: Credential Type. :param str value: Credential Value. """ if temporary is not None: pulumi.set(__self__, "temporary", temporary) if type is not None: pulumi.set(__self__, "type", type) if value is not None: pulumi.set(__self__, "value", value) @property @pulumi.getter def temporary(self) -> Optional[bool]: """ True if this credential object is temporary. """ return pulumi.get(self, "temporary") @property @pulumi.getter def type(self) -> Optional[str]: """ Credential Type. """ return pulumi.get(self, "type") @property @pulumi.getter def value(self) -> Optional[str]: """ Credential Value. """ return pulumi.get(self, "value") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakUserSpecUserFederatedIdentities(dict): def __init__(__self__, *, identity_provider: Optional[str] = None, user_id: Optional[str] = None, user_name: Optional[str] = None): """ :param str identity_provider: Federated Identity Provider. :param str user_id: Federated Identity User ID. :param str user_name: Federated Identity User Name. """ if identity_provider is not None: pulumi.set(__self__, "identity_provider", identity_provider) if user_id is not None: pulumi.set(__self__, "user_id", user_id) if user_name is not None: pulumi.set(__self__, "user_name", user_name) @property @pulumi.getter(name="identityProvider") def identity_provider(self) -> Optional[str]: """ Federated Identity Provider. """ return pulumi.get(self, "identity_provider") @property @pulumi.getter(name="userId") def user_id(self) -> Optional[str]: """ Federated Identity User ID. """ return pulumi.get(self, "user_id") @property @pulumi.getter(name="userName") def user_name(self) -> Optional[str]: """ Federated Identity User Name. """ return pulumi.get(self, "user_name") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop @pulumi.output_type class KeycloakUserStatus(dict): """ KeycloakUserStatus defines the observed state of KeycloakUser. """ def __init__(__self__, *, message: str, phase: str): """ KeycloakUserStatus defines the observed state of KeycloakUser. :param str message: Human-readable message indicating details about current operator phase or error. :param str phase: Current phase of the operator. """ pulumi.set(__self__, "message", message) pulumi.set(__self__, "phase", phase) @property @pulumi.getter def message(self) -> str: """ Human-readable message indicating details about current operator phase or error. """ return pulumi.get(self, "message") @property @pulumi.getter def phase(self) -> str: """ Current phase of the operator. """ return pulumi.get(self, "phase") def _translate_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
40.08212
638
0.645313
14,415
129,345
5.596393
0.040444
0.024655
0.035613
0.05205
0.857906
0.833127
0.81581
0.788179
0.779986
0.768916
0
0.000461
0.262329
129,345
3,226
639
40.094544
0.845033
0.325679
0
0.802602
1
0
0.135838
0.06598
0
0
0
0.00248
0
1
0.17138
false
0
0.003394
0.023756
0.346154
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
cb7564450627b8e178cddf0547e6a56a08190805
3,473
py
Python
netbox/ipam/migrations/0056_standardize_id_fields.py
cybarox/netbox
ea197eff5f4fe925bb354d1375912decd81752bd
[ "Apache-2.0" ]
null
null
null
netbox/ipam/migrations/0056_standardize_id_fields.py
cybarox/netbox
ea197eff5f4fe925bb354d1375912decd81752bd
[ "Apache-2.0" ]
null
null
null
netbox/ipam/migrations/0056_standardize_id_fields.py
cybarox/netbox
ea197eff5f4fe925bb354d1375912decd81752bd
[ "Apache-2.0" ]
null
null
null
from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('ipam', '0055_servicetemplate'), ] operations = [ # Model IDs migrations.AlterField( model_name='aggregate', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='asn', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='fhrpgroup', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='fhrpgroupassignment', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='ipaddress', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='iprange', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='prefix', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='rir', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='role', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='routetarget', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='service', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='servicetemplate', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='vlan', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='vlangroup', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), migrations.AlterField( model_name='vrf', name='id', field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False), ), # GFK IDs migrations.AlterField( model_name='fhrpgroupassignment', name='interface_id', field=models.PositiveBigIntegerField(), ), migrations.AlterField( model_name='ipaddress', name='assigned_object_id', field=models.PositiveBigIntegerField(blank=True, null=True), ), ]
34.73
92
0.580478
316
3,473
6.218354
0.151899
0.173028
0.216285
0.250891
0.836132
0.818321
0.756234
0.756234
0.756234
0.756234
0
0.001663
0.307515
3,473
99
93
35.080808
0.815385
0.004895
0
0.73913
0
0
0.066609
0
0
0
0
0
0
1
0
false
0.021739
0.01087
0
0.043478
0
0
0
0
null
0
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
9
38116738d8045592107d3e42e173ba9ef95f4212
19,855
py
Python
zaqar_tempest_plugin/services/messaging/json/messaging_client.py
openstack/zaqar-tempest-plugin
3813c99e501a1f67a9cbde751cf71747a61f6786
[ "Apache-2.0" ]
5
2017-10-31T13:38:16.000Z
2019-01-28T22:14:03.000Z
zaqar_tempest_plugin/services/messaging/json/messaging_client.py
openstack/zaqar-tempest-plugin
3813c99e501a1f67a9cbde751cf71747a61f6786
[ "Apache-2.0" ]
null
null
null
zaqar_tempest_plugin/services/messaging/json/messaging_client.py
openstack/zaqar-tempest-plugin
3813c99e501a1f67a9cbde751cf71747a61f6786
[ "Apache-2.0" ]
null
null
null
# Copyright (c) 2014 Rackspace, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. import urllib from oslo_utils import uuidutils from oslo_serialization import jsonutils as json from tempest.lib.common import rest_client from zaqar_tempest_plugin.api_schema.response.v1 \ import queues as v1schema from zaqar_tempest_plugin.api_schema.response.v1_1 \ import queues as v11schema from zaqar_tempest_plugin.api_schema.response.v2 \ import queues as v2schema class MessagingClient(rest_client.RestClient): def __init__(self, auth_provider, service, region, **kwargs): super(MessagingClient, self).__init__( auth_provider, service, region, **kwargs) self.version = '1' self.uri_prefix = 'v{0}'.format(self.version) client_id = uuidutils.generate_uuid(dashed=False) self.headers = {'Client-ID': client_id} class V1MessagingClient(MessagingClient): def __init__(self, auth_provider, service, region, **kwargs): super(V1MessagingClient, self).__init__( auth_provider, service, region, **kwargs) self.version = '1' def list_queues(self): uri = '{0}/queues'.format(self.uri_prefix) resp, body = self.get(uri) if resp['status'] != '204': body = json.loads(body) self.validate_response(v1schema.list_queues, resp, body) return resp, body def create_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.put(uri, body=None) self.expected_success(201, resp.status) return resp, body def show_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.get(uri) self.expected_success(204, resp.status) return resp, body def head_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.head(uri) self.expected_success(204, resp.status) return resp, body def delete_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.delete(uri) self.expected_success(204, resp.status) return resp, body def show_queue_stats(self, queue_name): uri = '{0}/queues/{1}/stats'.format(self.uri_prefix, queue_name) resp, body = self.get(uri) body = json.loads(body) self.validate_response(v1schema.queue_stats, resp, body) return resp, body def show_queue_metadata(self, queue_name): uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name) resp, body = self.get(uri) self.expected_success(200, resp.status) body = json.loads(body) return resp, body def set_queue_metadata(self, queue_name, rbody): uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name) resp, body = self.put(uri, body=json.dumps(rbody)) self.expected_success(204, resp.status) return resp, body def post_messages(self, queue_name, rbody): uri = '{0}/queues/{1}/messages'.format(self.uri_prefix, queue_name) resp, body = self.post(uri, body=json.dumps(rbody), extra_headers=True, headers=self.headers) body = json.loads(body) self.validate_response(v1schema.post_messages, resp, body) return resp, body def list_messages(self, queue_name): uri = '{0}/queues/{1}/messages?echo=True'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, extra_headers=True, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v1schema.list_messages, resp, body) return resp, body def show_single_message(self, message_uri): resp, body = self.get(message_uri, extra_headers=True, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v1schema.get_single_message, resp, body) return resp, body def show_multiple_messages(self, message_uri): resp, body = self.get(message_uri, extra_headers=True, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v1schema.get_multiple_messages, resp, body) return resp, body def delete_messages(self, message_uri): resp, body = self.delete(message_uri) self.expected_success(204, resp.status) return resp, body def post_claims(self, queue_name, rbody, url_params=False): uri = '{0}/queues/{1}/claims'.format(self.uri_prefix, queue_name) if url_params: uri += '?%s' % urllib.parse.urlencode(url_params) resp, body = self.post(uri, body=json.dumps(rbody), extra_headers=True, headers=self.headers) body = json.loads(body) self.validate_response(v1schema.claim_messages, resp, body) return resp, body def query_claim(self, claim_uri): resp, body = self.get(claim_uri) if resp['status'] != '204': body = json.loads(body) self.validate_response(v1schema.query_claim, resp, body) return resp, body def update_claim(self, claim_uri, rbody): resp, body = self.patch(claim_uri, body=json.dumps(rbody)) self.expected_success(204, resp.status) return resp, body def delete_claim(self, claim_uri): resp, body = self.delete(claim_uri) self.expected_success(204, resp.status) return resp, body class V11MessagingClient(MessagingClient): def __init__(self, auth_provider, service, region, **kwargs): super(V11MessagingClient, self).__init__( auth_provider, service, region, **kwargs) self.version = '1.1' self.uri_prefix = 'v{0}'.format(self.version) def list_queues(self): uri = '{0}/queues'.format(self.uri_prefix) resp, body = self.get(uri, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v11schema.list_queues, resp, body) return resp, body def create_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.put(uri, body=None, headers=self.headers) self.expected_success(201, resp.status) return resp, body def show_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, headers=self.headers) self.expected_success(200, resp.status) return resp, body def delete_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.delete(uri, headers=self.headers) self.expected_success(204, resp.status) return resp, body def show_queue_stats(self, queue_name): uri = '{0}/queues/{1}/stats'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, headers=self.headers) body = json.loads(body) self.validate_response(v11schema.queue_stats, resp, body) return resp, body def show_queue_metadata(self, queue_name): uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, headers=self.headers) self.expected_success(200, resp.status) body = json.loads(body) return resp, body def set_queue_metadata(self, queue_name, rbody): uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name) resp, body = self.put(uri, body=json.dumps(rbody), headers=self.headers) self.expected_success(204, resp.status) return resp, body def post_messages(self, queue_name, rbody): uri = '{0}/queues/{1}/messages'.format(self.uri_prefix, queue_name) resp, body = self.post(uri, body=json.dumps(rbody), extra_headers=True, headers=self.headers) body = json.loads(body) self.validate_response(v11schema.post_messages, resp, body) return resp, body def list_messages(self, queue_name): uri = '{0}/queues/{1}/messages?echo=True'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, extra_headers=True, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v11schema.list_messages, resp, body) return resp, body def show_single_message(self, message_uri): resp, body = self.get(message_uri, extra_headers=True, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v11schema.get_single_message, resp, body) return resp, body def show_multiple_messages(self, message_uri): resp, body = self.get(message_uri, extra_headers=True, headers=self.headers) if resp['status'] != '404': body = json.loads(body) self.validate_response(v11schema.get_multiple_messages, resp, body) return resp, body def delete_messages(self, message_uri): resp, body = self.delete(message_uri, headers=self.headers) self.expected_success(204, resp.status) return resp, body def post_claims(self, queue_name, rbody, url_params=False): uri = '{0}/queues/{1}/claims'.format(self.uri_prefix, queue_name) if url_params: uri += '?%s' % urllib.parse.urlencode(url_params) resp, body = self.post(uri, body=json.dumps(rbody), extra_headers=True, headers=self.headers) body = json.loads(body) self.validate_response(v11schema.claim_messages, resp, body) return resp, body def query_claim(self, claim_uri): resp, body = self.get(claim_uri, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v11schema.query_claim, resp, body) return resp, body def update_claim(self, claim_uri, rbody): resp, body = self.patch(claim_uri, body=json.dumps(rbody), headers=self.headers) self.expected_success(204, resp.status) return resp, body def delete_claim(self, claim_uri): resp, body = self.delete(claim_uri, headers=self.headers) self.expected_success(204, resp.status) return resp, body class V2MessagingClient(MessagingClient): def __init__(self, auth_provider, service, region, **kwargs): super(V2MessagingClient, self).__init__( auth_provider, service, region, **kwargs) self.version = '2' self.uri_prefix = 'v{0}'.format(self.version) def list_queues(self, url_params=False): uri = '{0}/queues'.format(self.uri_prefix) if url_params: uri += '?%s' % urllib.parse.urlencode(url_params) resp, body = self.get(uri, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v2schema.list_queues, resp, body) return resp, body def create_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.put(uri, body=None, headers=self.headers) self.expected_success(201, resp.status) return resp, body def show_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, headers=self.headers) self.expected_success(200, resp.status) return resp, body def delete_queue(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.delete(uri, headers=self.headers) self.expected_success(204, resp.status) return resp, body def purge_queue(self, queue_name, resource=None): uri = '{0}/queues/{1}/purge'.format(self.uri_prefix, queue_name) rbody = {"resource_types": ["messages", "subscriptions"]} if resource: rbody = {"resource_types": resource} resp, body = self.post(uri, body=json.dumps(rbody), headers=self.headers) self.expected_success(204, resp.status) return resp, body def show_queue_stats(self, queue_name): uri = '{0}/queues/{1}/stats'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, headers=self.headers) body = json.loads(body) self.validate_response(v2schema.queue_stats, resp, body) return resp, body def show_queue_metadata(self, queue_name): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, headers=self.headers) self.expected_success(200, resp.status) body = json.loads(body) return resp, body def set_queue_metadata(self, queue_name, rbody): uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name) headers = self.headers.copy() headers['Content-Type'] =\ 'application/openstack-messaging-v2.0-json-patch' resp, body = self.patch(uri, body=json.dumps(rbody), headers=headers) self.expected_success(200, resp.status) return resp, body def post_messages(self, queue_name, rbody): uri = '{0}/queues/{1}/messages'.format(self.uri_prefix, queue_name) resp, body = self.post(uri, body=json.dumps(rbody), extra_headers=True, headers=self.headers) body = json.loads(body) self.validate_response(v2schema.post_messages, resp, body) return resp, body def list_messages(self, queue_name): uri = '{0}/queues/{1}/messages?echo=True'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, extra_headers=True, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v2schema.list_messages, resp, body) return resp, body def show_single_message(self, message_uri): resp, body = self.get(message_uri, extra_headers=True, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v2schema.get_single_message, resp, body) return resp, body def show_multiple_messages(self, message_uri): resp, body = self.get(message_uri, extra_headers=True, headers=self.headers) if resp['status'] != '404': body = json.loads(body) self.validate_response(v2schema.get_multiple_messages, resp, body) return resp, body def delete_messages(self, message_uri): resp, body = self.delete(message_uri, headers=self.headers) self.expected_success(204, resp.status) return resp, body def post_claims(self, queue_name, rbody, url_params=False): uri = '{0}/queues/{1}/claims'.format(self.uri_prefix, queue_name) if url_params: uri += '?%s' % urllib.parse.urlencode(url_params) resp, body = self.post(uri, body=json.dumps(rbody), extra_headers=True, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v2schema.claim_messages, resp, body) return resp, body def query_claim(self, claim_uri): resp, body = self.get(claim_uri, headers=self.headers) if resp['status'] != '204': body = json.loads(body) self.validate_response(v2schema.query_claim, resp, body) return resp, body def update_claim(self, claim_uri, rbody): resp, body = self.patch(claim_uri, body=json.dumps(rbody), headers=self.headers) self.expected_success(204, resp.status) return resp, body def delete_claim(self, claim_uri): resp, body = self.delete(claim_uri, headers=self.headers) self.expected_success(204, resp.status) return resp, body def create_subscription(self, queue_name, rbody): uri = '{0}/queues/{1}/subscriptions'.format(self.uri_prefix, queue_name) resp, body = self.post(uri, body=json.dumps(rbody), extra_headers=True, headers=self.headers) body = json.loads(body) self.validate_response(v2schema.create_subscription, resp, body) return resp, body def delete_subscription(self, queue_name, subscription_id): uri = '{0}/queues/{1}/subscriptions/{2}'.format(self.uri_prefix, queue_name, subscription_id) resp, body = self.delete(uri, headers=self.headers) return resp, body def list_subscription(self, queue_name): uri = '{0}/queues/{1}/subscriptions'.format(self.uri_prefix, queue_name) resp, body = self.get(uri, headers=self.headers) body = json.loads(body) self.validate_response(v2schema.list_subscriptions, resp, body) return resp, body def show_subscription(self, queue_name, subscription_id): uri = '{0}/queues/{1}/subscriptions/{2}'.format(self.uri_prefix, queue_name, subscription_id) resp, body = self.get(uri, headers=self.headers) body = json.loads(body) self.validate_response(v2schema.show_single_subscription, resp, body) return resp, body def update_subscription(self, queue_name, subscription_id, rbody): uri = '{0}/queues/{1}/subscriptions/{2}'.format(self.uri_prefix, queue_name, subscription_id) resp, body = self.patch(uri, body=json.dumps(rbody), headers=self.headers) return resp, body
38.779297
77
0.59592
2,380
19,855
4.805882
0.071849
0.095821
0.057702
0.077286
0.901207
0.90007
0.890103
0.872967
0.856968
0.842368
0
0.019151
0.292571
19,855
511
78
38.855186
0.795173
0.027953
0
0.813131
0
0
0.054135
0.027068
0
0
0
0
0
1
0.14899
false
0
0.017677
0
0.315657
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
3836019055aa4cd174d05be7e971ed828f0eed54
12,757
py
Python
CIFAR10/more_attack.py
haiphanNJIT/StoBatch
08a6505b1fad609850de6c6bbaaa54c0200f0adb
[ "MIT" ]
6
2020-06-13T17:05:52.000Z
2021-12-02T12:41:39.000Z
MNIST/more_attack.py
haiphanNJIT/StoBatch
08a6505b1fad609850de6c6bbaaa54c0200f0adb
[ "MIT" ]
null
null
null
MNIST/more_attack.py
haiphanNJIT/StoBatch
08a6505b1fad609850de6c6bbaaa54c0200f0adb
[ "MIT" ]
1
2021-04-05T11:20:22.000Z
2021-04-05T11:20:22.000Z
import numpy as np from six.moves import xrange import tensorflow as tf from cleverhans.attacks_tf import fgm, fgsm from build_utils import batch_adv def model_loss(y, model, mean=True): """ FROM cleverhans/utils_tf Define loss of TF graph :param y: correct labels :param model: output of the model :param mean: boolean indicating whether should return mean of loss or vector of losses for each input of the batch :return: return mean of loss if True, otherwise return vector with per sample loss """ op = model.op if "softmax" in str(op).lower(): logits, = op.inputs else: logits = model out = tf.nn.softmax_cross_entropy_with_logits(logits=logits, labels=y) if mean: out = tf.reduce_mean(out) return out def fgm_pre_computed_grad(x, grad, eps=0.3, ord=np.inf, clip_min=None, clip_max=None, targeted=False): """ TensorFlow implementation of the Fast Gradient Method using pre computed gradients. :param x: input :param grad: pre-computed gradients for x on the pre-trained model (use negative (flipped) loss for the gradient if targeted) :param eps: the epsilon (input variation parameter) :param ord: (optional) Order of the norm (mimics NumPy). Possible values: np.inf, 1 or 2. :param clip_min: Minimum float value for adversarial example components :param clip_max: Maximum float value for adversarial example components :param targeted: Is the attack targeted or untargeted? Untargeted, the default, will try to make the label incorrect. Targeted will instead try to move in the direction of being more like y. :return: a tensor for the adversarial example """ if ord == np.inf: # Take sign of gradient normalized_grad = tf.sign(grad) # The following line should not change the numerical results. # It applies only because `normalized_grad` is the output of # a `sign` op, which has zero derivative anyway. # It should not be applied for the other norms, where the # perturbation has a non-zero derivative. normalized_grad = tf.stop_gradient(normalized_grad) elif ord == 1: red_ind = list(xrange(1, len(x.get_shape()))) normalized_grad = grad / tf.reduce_sum(tf.abs(grad), reduction_indices=red_ind, keep_dims=True) elif ord == 2: red_ind = list(xrange(1, len(x.get_shape()))) square = tf.reduce_sum(tf.square(grad), reduction_indices=red_ind, keep_dims=True) normalized_grad = grad / tf.sqrt(square) else: raise NotImplementedError("Only L-inf, L1 and L2 norms are " "currently implemented.") # Multiply by constant epsilon scaled_grad = eps * normalized_grad # Add perturbation to original example to obtain adversarial example adv_x = x + scaled_grad # If clipping is needed, reset all values outside of [clip_min, clip_max] if (clip_min is not None) and (clip_max is not None): adv_x = tf.clip_by_value(adv_x, clip_min, clip_max) return adv_x def rand_fgm(sess, x, logits, y=None, eps=0.3, ord=np.inf, rand_eps=0.3, rand_alpha=0.05, clip_min=None, clip_max=None, targeted=False): """ TensorFlow implementation of the Fast Gradient Method. :param x: the input placeholder :param preds: the model's output tensor (the attack expects the probabilities, i.e., the output of the softmax) :param y: (optional) A placeholder for the model labels. If targeted is true, then provide the target label. Otherwise, only provide this parameter if you'd like to use true labels when crafting adversarial samples. Otherwise, model predictions are used as labels to avoid the "label leaking" effect (explained in this paper: https://arxiv.org/abs/1611.01236). Default is None. Labels should be one-hot-encoded. :param eps: the epsilon (input variation parameter) :param ord: (optional) Order of the norm (mimics NumPy). Possible values: np.inf, 1 or 2. :param clip_min: Minimum float value for adversarial example components :param clip_max: Maximum float value for adversarial example components :param targeted: Is the attack targeted or untargeted? Untargeted, the default, will try to make the label incorrect. Targeted will instead try to move in the direction of being more like y. :return: a tensor for the adversarial example """ x_rand = x + rand_alpha * tf.sign(tf.random_normal(shape=tf.get_shape(x), mean=0.0, stddev=1.0)) if y is None: # Using model predictions as ground truth to avoid label leaking preds_max = tf.reduce_max(preds, 1, keep_dims=True) y = tf.to_float(tf.equal(preds, preds_max)) y = tf.stop_gradient(y) y = y / tf.reduce_sum(y, 1, keep_dims=True) # Compute loss loss = utils_tf.model_loss(y, preds, mean=False) if targeted: loss = -loss # Define gradient of loss wrt input grad, = tf.gradients(loss, x) if ord == np.inf: # Take sign of gradient normalized_grad = tf.sign(grad) # The following line should not change the numerical results. # It applies only because `normalized_grad` is the output of # a `sign` op, which has zero derivative anyway. # It should not be applied for the other norms, where the # perturbation has a non-zero derivative. normalized_grad = tf.stop_gradient(normalized_grad) elif ord == 1: red_ind = list(xrange(1, len(x.get_shape()))) normalized_grad = grad / tf.reduce_sum(tf.abs(grad), reduction_indices=red_ind, keep_dims=True) elif ord == 2: red_ind = list(xrange(1, len(x.get_shape()))) square = tf.reduce_sum(tf.square(grad), reduction_indices=red_ind, keep_dims=True) normalized_grad = grad / tf.sqrt(square) else: raise NotImplementedError("Only L-inf, L1 and L2 norms are " "currently implemented.") # Multiply by constant epsilon scaled_grad = eps * normalized_grad # Add perturbation to original example to obtain adversarial example adv_x = x + scaled_grad # If clipping is needed, reset all values outside of [clip_min, clip_max] if (clip_min is not None) and (clip_max is not None): adv_x = tf.clip_by_value(adv_x, clip_min, clip_max) return adv_x def iter_fgsm(sess, x_input_t, labels_t, x_input, labels, batch_size, preds_t, target_labels_t, steps, total_eps, step_eps, clip_min=0.0, clip_max=1.0, ord=np.inf, targeted=False): """ I-FGSM attack. This function directly generate adv inputs """ eta_t = fgm(x_input_t, preds_t, y=target_labels_t, eps=step_eps, ord=ord, clip_min=clip_min, clip_max=clip_max, targeted=targeted) - x_input_t if ord == np.inf: eta_t = tf.clip_by_value(eta_t, -total_eps, total_eps) elif ord in [1, 2]: reduc_ind = list(xrange(1, len(tf.shape(eta_t)))) if ord == 1: norm = tf.reduce_sum(tf.abs(eta_t), reduction_indices=reduc_ind, keep_dims=True) elif ord == 2: norm = tf.sqrt(tf.reduce_sum(tf.square(eta_t), reduction_indices=reduc_ind, keep_dims=True)) eta_t = eta_t * total_eps / norm x_adv_t = x_input_t + eta_t x_adv = x_input for i in range(steps): x_adv = batch_adv(sess, x_adv_t, x_input_t, labels_t, x_adv, labels, batch_size) return adv_x def iter_fgsm_t(x_input_t, preds_t, target_labels_t, steps, total_eps, step_eps, clip_min=0.0, clip_max=1.0, ord=np.inf, targeted=False): """ I-FGSM attack. """ eta_t = fgm(x_input_t, preds_t, y=target_labels_t, eps=step_eps, ord=ord, clip_min=clip_min, clip_max=clip_max, targeted=targeted) - x_input_t if ord == np.inf: eta_t = tf.clip_by_value(eta_t, -total_eps, total_eps) elif ord in [1, 2]: reduc_ind = list(xrange(1, len(tf.shape(eta_t)))) if ord == 1: norm = tf.reduce_sum(tf.abs(eta_t), reduction_indices=reduc_ind, keep_dims=True) elif ord == 2: norm = tf.sqrt(tf.reduce_sum(tf.square(eta_t), reduction_indices=reduc_ind, keep_dims=True)) eta_t = eta_t * total_eps / norm x_adv_t = x_input_t + eta_t return x_adv_t def _fgm(x, preds, y=None, eps=0.3, ord=np.inf, clip_min=None, clip_max=None, targeted=False): """ TensorFlow implementation of the Fast Gradient Method. :param x: the input placeholder :param preds: the model's output tensor (the attack expects the probabilities, i.e., the output of the softmax) :param y: (optional) A placeholder for the model labels. If targeted is true, then provide the target label. Otherwise, only provide this parameter if you'd like to use true labels when crafting adversarial samples. Otherwise, model predictions are used as labels to avoid the "label leaking" effect (explained in this paper: https://arxiv.org/abs/1611.01236). Default is None. Labels should be one-hot-encoded. :param eps: the epsilon (input variation parameter) :param ord: (optional) Order of the norm (mimics NumPy). Possible values: np.inf, 1 or 2. :param clip_min: Minimum float value for adversarial example components :param clip_max: Maximum float value for adversarial example components :param targeted: Is the attack targeted or untargeted? Untargeted, the default, will try to make the label incorrect. Targeted will instead try to move in the direction of being more like y. :return: a tensor for the adversarial example """ if y is None: # Using model predictions as ground truth to avoid label leaking preds_max = tf.reduce_max(preds, 1, keep_dims=True) y = tf.to_float(tf.equal(preds, preds_max)) y = tf.stop_gradient(y) y = y / tf.reduce_sum(y, 1, keep_dims=True) # Compute loss loss = utils_tf.model_loss(y, preds, mean=False) if targeted: loss = -loss # Define gradient of loss wrt input grad, = tf.gradients(loss, x) if ord == np.inf: # Take sign of gradient normalized_grad = tf.sign(grad) # The following line should not change the numerical results. # It applies only because `normalized_grad` is the output of # a `sign` op, which has zero derivative anyway. # It should not be applied for the other norms, where the # perturbation has a non-zero derivative. normalized_grad = tf.stop_gradient(normalized_grad) elif ord == 1: red_ind = list(xrange(1, len(x.get_shape()))) normalized_grad = grad / tf.reduce_sum(tf.abs(grad), reduction_indices=red_ind, keep_dims=True) elif ord == 2: red_ind = list(xrange(1, len(x.get_shape()))) square = tf.reduce_sum(tf.square(grad), reduction_indices=red_ind, keep_dims=True) normalized_grad = grad / tf.sqrt(square) else: raise NotImplementedError("Only L-inf, L1 and L2 norms are " "currently implemented.") # Multiply by constant epsilon scaled_grad = eps * normalized_grad # Add perturbation to original example to obtain adversarial example adv_x = x + scaled_grad # If clipping is needed, reset all values outside of [clip_min, clip_max] if (clip_min is not None) and (clip_max is not None): adv_x = tf.clip_by_value(adv_x, clip_min, clip_max) return adv_x
42.38206
100
0.612997
1,780
12,757
4.241011
0.134831
0.019473
0.022255
0.017221
0.874156
0.873228
0.86475
0.86475
0.861704
0.861704
0
0.008961
0.308928
12,757
300
101
42.523333
0.847323
0.418907
0
0.818182
0
0
0.023972
0
0
0
0
0
0
1
0.038961
false
0
0.032468
0
0.11039
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
69944ec5508df035d8cb6588b4a276d892b033bc
51,749
py
Python
server/config/mv_fantasy/templates.py
AustralianDisabilityLimited/MultiversePlatform
7e1aad33d48b9e47f3db2ca638cb57592336ddb7
[ "MIT" ]
33
2015-02-16T02:52:08.000Z
2022-02-18T08:46:32.000Z
server/config/mv_fantasy/templates.py
bensku/MultiversePlatform
7e1aad33d48b9e47f3db2ca638cb57592336ddb7
[ "MIT" ]
1
2017-09-09T18:50:23.000Z
2020-12-29T18:13:56.000Z
server/config/mv_fantasy/templates.py
bensku/MultiversePlatform
7e1aad33d48b9e47f3db2ca638cb57592336ddb7
[ "MIT" ]
31
2015-02-07T16:20:24.000Z
2022-02-23T15:02:43.000Z
# # The Multiverse Platform is made available under the MIT License. # # Copyright (c) 2012 The Multiverse Foundation # # Permission is hereby granted, free of charge, to any person # obtaining a copy of this software and associated documentation # files (the "Software"), to deal in the Software without restriction, # including without limitation the rights to use, copy, modify, # merge, publish, distribute, sublicense, and/or sell copies # of the Software, and to permit persons to whom the Software # is furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be # included in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, # EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES # OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT # HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, # WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE # OR OTHER DEALINGS IN THE SOFTWARE. # # from multiverse.mars import * from multiverse.mars.objects import * from multiverse.mars.core import * from multiverse.mars.events import * from multiverse.mars.util import * from multiverse.mars.plugins import * from multiverse.server.plugins import * from multiverse.server.math import * from multiverse.server.events import * from multiverse.server.objects import * from multiverse.server.engine import * from java.lang import * True=1 False=0 class TemplateHook(EnginePlugin.PluginActivateHook): def activate(self): ############################################################ # # DisplayContexts # ############################################################ human_female_base_DC = DisplayContext("human_female.mesh", True) human_female_base_DC.addSubmesh(DisplayContext.Submesh("bodyShape-lib.0", "human_female.skin_material")) human_female_base_DC.addSubmesh(DisplayContext.Submesh("head_aShape-lib.0", "human_female.head_a_material")) human_female_leather_pantsDC = DisplayContext() human_female_leather_pantsDC.setMeshFile("human_female.mesh") human_female_leather_pantsDC.addSubmesh(DisplayContext.Submesh("leather_a_pantsShape-lib.0", "human_female.leather_a_material")) human_female_leather_pantsDC.addSubmesh(DisplayContext.Submesh("leather_a_beltShape-lib.0", "human_female.leather_a_material")) human_female_leather_tunicDC = DisplayContext() human_female_leather_tunicDC.setMeshFile("human_female.mesh") human_female_leather_tunicDC.addSubmesh(DisplayContext.Submesh("leather_a_tunicShape-lib.0", "human_female.leather_a_material")) human_female_leather_bootsDC = DisplayContext() human_female_leather_bootsDC.setMeshFile("human_female.mesh") human_female_leather_bootsDC.addSubmesh(DisplayContext.Submesh("leather_a_bootsShape-lib.0", "human_female.leather_a_material")) human_female_leather_glovesDC = DisplayContext() human_female_leather_glovesDC.setMeshFile("human_female.mesh") human_female_leather_glovesDC.addSubmesh(DisplayContext.Submesh("leather_a_bracersShape-lib.0", "human_female.leather_a_material")) human_female_plate_tunicDC = DisplayContext() human_female_plate_tunicDC.setMeshFile("human_female.mesh") human_female_plate_tunicDC.addSubmesh(DisplayContext.Submesh("plate_b_tunicShape-lib.0", "human_female.plate_b_material")) zombie_base_DC = DisplayContext("zombie.mesh") zombie_base_DC.addSubmesh(DisplayContext.Submesh("Zombie_Body2-obj.0", "Zombie.Zombie_Body")) zombie_base_DC.addSubmesh(DisplayContext.Submesh("Zombie_Clothes2-obj.0", "Zombie.Zombie_Clothes")) orc_base_DC = DisplayContext("orc.mesh", True) human_female_ruth_base_DC = DisplayContext("human_female_ruth.mesh", True) human_female_ruth_base_DC.addSubmesh(DisplayContext.Submesh("human_female_body_ruth-mesh.0", "human_female_ruth.ruth_body_clothed_mat")) human_female_ruth_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth-mesh.0", "human_female_ruth.ruth_head_mat")) human_female_ruth_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth_hair-mesh.0", "human_female_ruth.ruth_hair_mat")) human_female_ruth_leather_pants_DC = DisplayContext("human_female_ruth.mesh") human_female_ruth_leather_pants_DC.addSubmesh(DisplayContext.Submesh("leather_armor_legs-mesh.0", "human_female_ruth.leather_armor_mat")) human_female_ruth_leather_pants_DC.addSubmesh(DisplayContext.Submesh("leather_armor_belt-mesh.0", "human_female_ruth.leather_armor_mat")) human_female_ruth_leather_tunic_DC = DisplayContext("human_female_ruth.mesh") human_female_ruth_leather_tunic_DC.addSubmesh(DisplayContext.Submesh("leather_armor_chest-mesh.0", "human_female_ruth.leather_armor_mat")) human_female_ruth_leather_boots_DC = DisplayContext("human_female_ruth.mesh") human_female_ruth_leather_boots_DC.addSubmesh(DisplayContext.Submesh("leather_armor_boots-mesh.0", "human_female_ruth.leather_armor_mat")) human_female_ruth_leather_gloves_DC = DisplayContext("human_female_ruth.mesh") human_female_ruth_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_rt-mesh.0", "human_female_ruth.leather_armor_mat")) human_female_ruth_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_lt-mesh.0", "human_female_ruth.leather_armor_mat")) human_male_DC = DisplayContext("human_male.mesh", True) human_male_DC.addSubmesh(DisplayContext.Submesh("head_bShape-lib.0", "human_male.head_a_material")) human_male_DC.addSubmesh(DisplayContext.Submesh("bodyShape-lib.0", "human_male.skin_material")) human_male_DC.addSubmesh(DisplayContext.Submesh("cloth_a_pantsShape-lib.0", "human_male.cloth_a_material")) human_male_DC.addSubmesh(DisplayContext.Submesh("cloth_a_bootsShape-lib.0", "human_male.cloth_a_material")) human_male_DC.addSubmesh(DisplayContext.Submesh("cloth_a_shirtShape-lib.0", "human_male.cloth_a_material")) female_player_01_base_DC = DisplayContext("human_female_fantasy.mesh", True) female_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_female_body-mesh.0", "human_female_fantasy.body_01_clothed_mat")) female_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_01-mesh.0", "human_female_fantasy.head_01_mat")) female_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_01_hair-mesh.0", "human_female_fantasy.head_01_hair_01_mat")) female_player_02_base_DC = DisplayContext("human_female_fantasy.mesh", True) female_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_female_body-mesh.0", "human_female_fantasy.body_02_clothed_mat")) female_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_02-mesh.0", "human_female_fantasy.head_02_mat")) female_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_02_hair_01-mesh.0", "human_female_fantasy.head_02_hair_01_mat")) female_player_leather_pants_DC = DisplayContext("human_female_fantasy.mesh") female_player_leather_pants_DC.addSubmesh(DisplayContext.Submesh("leather_armor_legs-mesh.0", "human_female_fantasy.leather_armor_mat")) female_player_leather_pants_DC.addSubmesh(DisplayContext.Submesh("leather_armor_belt-mesh.0", "human_female_fantasy.leather_armor_mat")) female_player_leather_tunic_DC = DisplayContext("human_female_fantasy.mesh") female_player_leather_tunic_DC.addSubmesh(DisplayContext.Submesh("leather_armor_chest-mesh.0", "human_female_fantasy.leather_armor_mat")) female_player_leather_tunic_DC.addSubmesh(DisplayContext.Submesh("leather_armor_jewels-mesh.0", "human_female_fantasy.leather_armor_mat")) female_player_leather_boots_DC = DisplayContext("human_female_fantasy.mesh") female_player_leather_boots_DC.addSubmesh(DisplayContext.Submesh("leather_armor_boot_lt-mesh.0", "human_female_fantasy.leather_armor_mat")) female_player_leather_boots_DC.addSubmesh(DisplayContext.Submesh("leather_armor_boot_rt-mesh.0", "human_female_fantasy.leather_armor_mat")) female_player_leather_gloves_DC = DisplayContext("human_female_fantasy.mesh") female_player_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_rt-mesh.0", "human_female_fantasy.leather_armor_mat")) female_player_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_lt-mesh.0", "human_female_fantasy.leather_armor_mat")) male_player_01_base_DC = DisplayContext("human_male_fantasy.mesh", True) male_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_male_body-mesh.0", "human_male_fantasy.human_male_body_01")) male_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_male_head_01-mesh.0", "human_male_fantasy.human_male_head_01")) male_player_01_base_DC.addSubmesh(DisplayContext.Submesh("male_head_01_hair_01-mesh.0", "human_male_fantasy.human_male_head_01_hair_01")) male_player_02_base_DC = DisplayContext("human_male_fantasy.mesh", True) male_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_male_body-mesh.0", "human_male_fantasy.human_male_body_02")) male_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_male_head_02-mesh.0", "human_male_fantasy.human_male_head_02")) male_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_male_02_hair_01-mesh.0", "human_male_fantasy.human_male_head_02_hair_01")) male_player_leather_pants_DC = DisplayContext("human_male_fantasy.mesh") male_player_leather_pants_DC.addSubmesh(DisplayContext.Submesh("male_leather_b_legs-mesh.0", "human_male_fantasy.human_male_armor_leather_b")) male_player_leather_tunic_DC = DisplayContext("human_male_fantasy.mesh") male_player_leather_tunic_DC.addSubmesh(DisplayContext.Submesh("male_leather_b_chest-mesh.0", "human_male_fantasy.human_male_armor_leather_b")) male_player_leather_boots_DC = DisplayContext("human_male_fantasy.mesh") male_player_leather_boots_DC.addSubmesh(DisplayContext.Submesh("male_leather_b_feet-mesh.0", "human_male_fantasy.human_male_armor_leather_b")) male_player_leather_gloves_DC = DisplayContext("human_male_fantasy.mesh") male_player_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("male_leather_b_hands-mesh.0", "human_male_fantasy.human_male_armor_leather_b")) hilldale_scout_DC = DisplayContext("human_female_ruth.mesh", True) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("human_female_body_ruth-mesh.0", "human_female_ruth.ruth_body_clothed_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth-mesh.0", "human_female_ruth.ruth_head_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth_hair-mesh.0", "human_female_ruth.ruth_hair_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_legs-mesh.0", "human_female_ruth.leather_armor_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_belt-mesh.0", "human_female_ruth.leather_armor_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_chest-mesh.0", "human_female_ruth.leather_armor_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_jewel-mesh.0", "human_female_ruth.leather_armor_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_boots-mesh.0", "human_female_ruth.leather_armor_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_rt-mesh.0", "human_female_ruth.leather_armor_mat")) hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_lt-mesh.0", "human_female_ruth.leather_armor_mat")) constable_dillon_DC = DisplayContext("human_female_ruth.mesh", True) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("human_female_body_ruth-mesh.0", "human_female_ruth.ruth_body_clothed_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth-mesh.0", "human_female_ruth.ruth_head_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxChest-mesh.0", "human_female_ruth.brax_armor_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxHelmetArmor-mesh.0", "human_female_ruth.brax_armor_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxLegs-mesh.0", "human_female_ruth.brax_armor_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxBracerLt-mesh.0", "human_female_ruth.brax_armor_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxBracerRt-mesh.0", "human_female_ruth.brax_armor_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxShoulderLt-mesh.0", "human_female_ruth.brax_armor_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxShoulderRt-mesh.0", "human_female_ruth.brax_armor_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxBelt-mesh.0", "human_female_ruth.brax_armor_mat")) constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxBoots-mesh.0", "human_female_ruth.brax_armor_mat")) brax_DC = DisplayContext("brax.mesh", True) brax_DC.addSubmesh(DisplayContext.Submesh("braxHeadShape.0", "brax.braxBodyMat")) brax_DC.addSubmesh(DisplayContext.Submesh("braxBodyShape.0", "brax.braxBodyMat")) brax_DC.addSubmesh(DisplayContext.Submesh("braxTailShape.0", "brax.braxBodyMat")) brax_DC.addSubmesh(DisplayContext.Submesh("brax_shell1Shape.0", "brax.braxBodyMat")) brax_DC.addSubmesh(DisplayContext.Submesh("brax_shell2Shape.0", "brax.braxBodyMat")) ############################################################# # # mob templates # ############################################################# defaultSlots = MarsEquipInfo("default") defaultSlots.addEquipSlot(MarsEquipSlot.PRIMARYWEAPON) # # Wolf Template # tmpl = Template("Wolf") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, DisplayContext("wolf.mesh", True)) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(InventoryClient.NAMESPACE, InventoryClient.TEMPL_ITEMS, "Wolf Skin") tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18)) tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18)) tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10)) tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10)) tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100)) tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100)) tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0)) tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability") tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect") tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10) tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5) ObjectManagerClient.registerTemplate(tmpl) # # Coyote Template # tmpl = Template("Coyote") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, DisplayContext("wolf.mesh", True)) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_SCALE, MVVector(0.75, 0.75, 0.75)) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18)) tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18)) tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10)) tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10)) tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100)) tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100)) tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0)) tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability") tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect") tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10) tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5) ObjectManagerClient.registerTemplate(tmpl) # # Crocodile Template # tmpl = Template("Crocodile") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, DisplayContext("crocodile.mesh", True)) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_RUN_THRESHOLD, Float(7000)) tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18)) tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18)) tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10)) tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10)) tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100)) tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100)) tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0)) tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability") tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect") tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10) tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5) ObjectManagerClient.registerTemplate(tmpl) # # Zombie Template # tmpl = Template("Zombie") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, zombie_base_DC) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True)) tmpl.put(InventoryClient.NAMESPACE, InventoryClient.TEMPL_ITEMS, "Zombie Dust") tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18)) tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18)) tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10)) tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10)) tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100)) tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100)) tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0)) tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability") tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect") tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10) tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5) ObjectManagerClient.registerTemplate(tmpl) # # Brax Template # tmpl = Template("Brax") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, brax_DC) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_RUN_THRESHOLD, Float(7000)) tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18)) tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18)) tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10)) tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10)) tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100)) tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100)) tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0)) tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability") tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect") tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10) tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5) ObjectManagerClient.registerTemplate(tmpl) # # Orc Warrior Template # tmpl = Template("Orc Warrior") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, orc_base_DC) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(InventoryClient.NAMESPACE, InventoryClient.TEMPL_ITEMS, "*sword10") tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True)) tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18)) tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18)) tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10)) tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10)) tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100)) tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100)) tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5)) tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0)) tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability") tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect") tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10) tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5) ObjectManagerClient.registerTemplate(tmpl) # # Hilldale Scout # tmpl = Template("Hilldale Scout") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, hilldale_scout_DC) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_ORIENT, Quaternion(0, 0.468, 0, 0.884)) tmpl.put(InventoryClient.NAMESPACE, InventoryClient.TEMPL_ITEMS, "*Leather Tunic; *Leather Pants; *Leather Boots; *sword4") ObjectManagerClient.registerTemplate(tmpl) # # Constable Dillon # tmpl = Template("Constable Dillon") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, constable_dillon_DC) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_ORIENT, Quaternion(0, 0.857, 0, -0.515)) ObjectManagerClient.registerTemplate(tmpl) # # Cyrus Blackfire # tmpl = Template("Cyrus Blackfire") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, human_male_DC) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_NAME, "Cyrus Blackfire") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_ORIENT, Quaternion(0, 0.857, 0, -0.515)) ObjectManagerClient.registerTemplate(tmpl) # # Weapon Templates # equipInfo = MarsEquipInfo("weapon") equipInfo.addEquipSlot(MarsEquipSlot.PRIMARYWEAPON) dc = DisplayContext("sword.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("Bronze Longsword") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_basic.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword1") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_broad.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword2") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_katar.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword3") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_katareen.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword4") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_leaf.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword5") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_pointy.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword6") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_sabre.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword7") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_serpent.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword8") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_short.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword9") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) dc = DisplayContext("FW_sword_human_stone.mesh") dc.setAttachInfo(DisplayState.IN_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dc.setAttachInfo(DisplayState.NON_COMBAT, MarsEquipSlot.PRIMARYWEAPON, MarsAttachSocket.PRIMARYWEAPON) dcMap = DCMap() dcMap.add(human_female_base_DC, dc) dcMap.add(orc_base_DC, dc) dcMap.add(zombie_base_DC, dc) dcMap.add(human_female_ruth_base_DC, dc) dcMap.add(female_player_01_base_DC, dc) dcMap.add(female_player_02_base_DC, dc) dcMap.add(male_player_01_base_DC, dc) dcMap.add(male_player_02_base_DC, dc) tmpl = Template("sword10") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\WEAPON_sword_A") tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(tmpl) # # Armor Templates # equipInfo = MarsEquipInfo("armor") equipInfo.addEquipSlot(MarsEquipSlot.CHEST) dcMap = DCMap() dcMap.add(human_female_base_DC, human_female_leather_tunicDC) dcMap.add(human_female_ruth_base_DC, human_female_ruth_leather_tunic_DC) dcMap.add(female_player_01_base_DC, female_player_leather_tunic_DC) dcMap.add(female_player_02_base_DC, female_player_leather_tunic_DC) dcMap.add(male_player_01_base_DC, male_player_leather_tunic_DC) dcMap.add(male_player_02_base_DC, male_player_leather_tunic_DC) item = Template("Leather Tunic") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ARMOR_leather_A_chest") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(item) equipInfo = MarsEquipInfo("pants") equipInfo.addEquipSlot(MarsEquipSlot.LEGS) dcMap = DCMap() dcMap.add(human_female_base_DC, human_female_leather_pantsDC) dcMap.add(human_female_ruth_base_DC, human_female_ruth_leather_pants_DC) dcMap.add(female_player_01_base_DC, female_player_leather_pants_DC) dcMap.add(female_player_02_base_DC, female_player_leather_pants_DC) dcMap.add(male_player_01_base_DC, male_player_leather_pants_DC) dcMap.add(male_player_02_base_DC, male_player_leather_pants_DC) item = Template("Leather Pants") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ARMOR_leather_A_legs") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(item) equipInfo = MarsEquipInfo("boots") equipInfo.addEquipSlot(MarsEquipSlot.FEET) dcMap = DCMap() dcMap.add(human_female_base_DC, human_female_leather_bootsDC) dcMap.add(human_female_ruth_base_DC, human_female_ruth_leather_boots_DC) dcMap.add(female_player_01_base_DC, female_player_leather_boots_DC) dcMap.add(female_player_02_base_DC, female_player_leather_boots_DC) dcMap.add(male_player_01_base_DC, male_player_leather_boots_DC) dcMap.add(male_player_02_base_DC, male_player_leather_boots_DC) item = Template("Leather Boots") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ARMOR_leather_A_feet") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(item) equipInfo = MarsEquipInfo("gloves") equipInfo.addEquipSlot(MarsEquipSlot.HANDS) dcMap = DCMap() dcMap.add(human_female_base_DC, human_female_leather_glovesDC) dcMap.add(human_female_ruth_base_DC, human_female_ruth_leather_gloves_DC) dcMap.add(female_player_01_base_DC, female_player_leather_gloves_DC) dcMap.add(female_player_02_base_DC, female_player_leather_gloves_DC) dcMap.add(male_player_01_base_DC, male_player_leather_gloves_DC) dcMap.add(male_player_02_base_DC, male_player_leather_gloves_DC) item = Template("Leather Gloves") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ARMOR_leather_A_hands") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook()) item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo) item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap) ObjectManagerClient.registerTemplate(item) # # Item Templates # item = Template("Wolf Skin") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\Icons\INV_hide_basic-pelt") ObjectManagerClient.registerTemplate(item) item = Template("Zombie Dust") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\Icons\INV_mined_ironore") ObjectManagerClient.registerTemplate(item) item = Template("Healing Potion") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ITEM_potion_A") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, AbilityActivateHook("heal potion")) ObjectManagerClient.registerTemplate(item) item = Template("Mana Potion") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ITEM_potion_A") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, AbilityActivateHook("restore mana potion")) ObjectManagerClient.registerTemplate(item) item = Template("Poison Potion") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ITEM_potion_A") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, AbilityActivateHook("poison potion")) ObjectManagerClient.registerTemplate(item) item = Template("Tome of Heal") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ITEM_book_C") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, AbilityActivateHook("teach self heal ability")) ObjectManagerClient.registerTemplate(item) item = Template("Tome of Fireball") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON, "Interface\FantasyWorldIcons\ITEM_book_A") item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, AbilityActivateHook("teach self fireball ability")) ObjectManagerClient.registerTemplate(item) # # Teleporter # tmpl = Template("Teleporter") tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, DisplayContext("tiny_cube.mesh")) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob) tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_NAME, "") ObjectManagerClient.registerTemplate(tmpl)
60.809636
121
0.66239
5,255
51,749
6.203045
0.06432
0.035433
0.02853
0.076449
0.88956
0.87413
0.864497
0.827561
0.803356
0.772034
0
0.012101
0.247889
51,749
850
122
60.881176
0.825416
0.027054
0
0.643972
0
0
0.158975
0.110327
0
0
0
0
0
0
null
null
0
0.017021
null
null
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
7
699509467b4776fdc0b97a22116737e46ae9cf57
21,902
py
Python
Spped Advisory.py
HossseinMoradi/Project4
dbdbbcd1ef8aded9b86bb23b91657e6f6501806e
[ "MIT" ]
null
null
null
Spped Advisory.py
HossseinMoradi/Project4
dbdbbcd1ef8aded9b86bb23b91657e6f6501806e
[ "MIT" ]
null
null
null
Spped Advisory.py
HossseinMoradi/Project4
dbdbbcd1ef8aded9b86bb23b91657e6f6501806e
[ "MIT" ]
null
null
null
import math import ast from math import e from decimal import * import time import random getcontext().prec = 28 # first we have defined functions to access signals and vehicles data def toList(NestedTuple): return list(map(toList, NestedTuple)) if isinstance(NestedTuple, (list, tuple)) else NestedTuple # we want to access vehicles information. In the first part of the below function, we determine the pieces of data that are collected. In the secoond part we determine that which types of vehicles are CV/special (i.e, sending information) def GetVissimDataVehicles(): global vehsAttributes global vehsAttNames vehsAttributesNames = ['No', 'VehType\No', 'Pos', 'VehType\No', 'Lane\Link','DesSpeed', 'Speed', 'DistanceToSigHead','InQueue'] vehsAttributes = toList(Vissim.Net.Vehicles.GetMultipleAttributes(vehsAttributesNames)) vehsAttNames = {} cnt = 0 for att in vehsAttributesNames: vehsAttNames.update({att: cnt}) cnt += 1 global vehTypesEquipped global vehTypesSpecial vehTypesAttributes = Vissim.Net.VehicleTypes.GetMultipleAttributes(['No', 'IsCV', 'IsSpecial']) vehTypesEquipped = [x[0] for x in vehTypesAttributes if x[1] == True] vehTypesSpecial = [x[0] for x in vehTypesAttributes if x[2] == True] def Signal(): #we define a user attributre to access SimSec Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('SimSec',Vissim.Net.Simulation.SimulationSecond) Seconds = Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycSec') SimSec = Vissim.Net.SignalControllers.ItemByKey(1).AttValue('SimSec') CLength = 60 GetVissimDataVehicles() deltaT=1 #we should correlate deltaT with the simulation resolution. In other words, number of simulation per second should be one here. Starting_time=0 Ending_time=1000 TimeNo=[] i=Starting_time k=0 while i< Ending_time: TimeNo.append(k) k+=1 i=i+deltaT # Here we define the following attributes: TimeUntilNextGreen, TimeUntilNextRed, CycleDuration for i in TimeNo: if SimSec > (i)*deltaT and SimSec <= (i+1)*deltaT: # this is anexample of a case where green time durations and cycles are varying at each time step if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('SigState')=='RED': if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart'): G1=random.randint(10,17) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenTimeDuration', G1) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenTimeDuration')) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart') < SimSec : Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')+ Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenTimeDuration')) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('SigState')=='GREEN': G1=max(random.randint(10,17),SimSec-Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart')) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenTimeDuration', G1) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextGreen', -1) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd')-SimSec) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('SigState')=='RED': if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart'): G2=random.randint(10,17) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('GreenTimeDuration', G2) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenTimeDuration')) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart') < SimSec : Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')+ Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenTimeDuration')) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('SigState')=='GREEN': G2=max(random.randint(10,17),SimSec-Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart')) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('GreenTimeDuration', G2) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextGreen', -1) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd')-SimSec) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('SigState')=='RED': if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart'): G3=random.randint(10,17) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('GreenTimeDuration', G3) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenTimeDuration')) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart') < SimSec : Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')+ Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenTimeDuration')) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('SigState')=='GREEN': G3=max(random.randint(10,17),SimSec-Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart')) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('GreenTimeDuration', G3) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextGreen', -1) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd')-SimSec) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('SigState')=='RED': if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart'): G4=random.randint(10,17) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('GreenTimeDuration', G4) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenTimeDuration')) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart') < SimSec : Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')+ Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenTimeDuration')) if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('SigState')=='GREEN': G4=max(random.randint(10,17),SimSec-Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart')) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('GreenTimeDuration', G4) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextGreen', -1) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd')-SimSec) Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('CycleDuration',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenTimeDuration')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenTimeDuration')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenTimeDuration')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenTimeDuration')+8) Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('CycleEnd',Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleStart')+Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')) if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd'): if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd')+2: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenStart', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd')+2) Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('CycleStart', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart')-2) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenEnd', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart') + Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenTimeDuration')) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('GreenStart',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd') + 2) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('GreenEnd',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart') + Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenTimeDuration')) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('GreenStart',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd') + 2) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('GreenEnd',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart') + Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenTimeDuration')) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('GreenStart',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd') + 2) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('GreenEnd',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart') + Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenTimeDuration')) # you have to make a use of simsec to define variable cycle time # To this end, we firs, define three attributes in the current signal controller with the names of CycleStart, CycleEnd, CycleDuration. SimSec = Vissim.Net.SignalControllers.ItemByKey(1).AttValue('SimSec') #we determine Initial values OF CYCLE START if SimSec<=1: Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('CycleStart', 0) # When we start the simulation, we determine that the signals are operating upon com script. if SimSec<=1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('ContrByCOM', True) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('ContrByCOM', True) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('ContrByCOM', True) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('ContrByCOM', True) Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenStart',2) if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart') - 1: if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd') - 1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'GREEN') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED') if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd') - 1: if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd') + 1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED') if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart') - 1: if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd') - 1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'GREEN') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED') if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd') - 1: if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd') + 1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED') if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart') - 1: if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd') - 1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'GREEN') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED') if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd') - 1: if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd') + 1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED') if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart') - 1: if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd') - 1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'GREEN') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED') if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd') - 1: Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED') Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED') def ChangeSpeed(): GetVissimDataVehicles() MinSpeed=5 MaxSpeed=76 if len(vehsAttributes) > 1: for vehAttributes in vehsAttributes: if vehAttributes[vehsAttNames['VehType\\No']] in vehTypesEquipped: # set easier variables of the current vehicle: DesSpeed = vehAttributes[vehsAttNames['DesSpeed']] Speed = vehAttributes[vehsAttNames['Speed']] DistanceToSigHead = vehAttributes[vehsAttNames['DistanceToSigHead']] # if the vehicle does not have a upcoming signal: set original desired speed if DistanceToSigHead <= 0: vehAttributes[vehsAttNames['DesSpeed']] = MaxSpeed continue # jump to next vehicle elif vehAttributes[vehsAttNames['Lane\Link']] == '1': if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('TimeUntilNextGreen')==-1: vehAttributes[vehsAttNames['DesSpeed']] = MaxSpeed if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('TimeUntilNextGreen')!=-1: vehAttributes[vehsAttNames['DesSpeed']] = MinSpeed vehicleNumDesiredSpeeds = [[x[vehsAttNames['DesSpeed']], x[vehsAttNames['Speed']]] for x in vehsAttributes] Vissim.Net.Vehicles.SetMultipleAttributes(('DesSpeed', 'Speed'), vehicleNumDesiredSpeeds)
82.338346
441
0.697516
2,333
21,902
6.546507
0.084869
0.142736
0.297911
0.401035
0.841616
0.841616
0.832122
0.8187
0.804099
0.797289
0
0.024041
0.162451
21,902
265
442
82.649057
0.808548
0.051913
0
0.278075
0
0
0.115152
0
0
0
0
0
0
0
null
null
0
0.032086
null
null
0
0
0
0
null
0
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
9
69b20446b0c37d914c58cb544ec08ec1b2212bcc
11,088
py
Python
skgaip/tspdb/tspdb/src/algorithms/pymf/nmf.py
danielsuo/toy_flood
471d3c4091d86d4a00fbf910937d4e60fdaf79a1
[ "MIT" ]
43
2019-12-10T00:05:51.000Z
2022-03-31T21:21:20.000Z
skgaip/tspdb/tspdb/src/algorithms/pymf/nmf.py
danielsuo/toy_flood
471d3c4091d86d4a00fbf910937d4e60fdaf79a1
[ "MIT" ]
5
2021-05-09T01:12:31.000Z
2022-03-29T17:34:15.000Z
skgaip/tspdb/tspdb/src/algorithms/pymf/nmf.py
danielsuo/toy_flood
471d3c4091d86d4a00fbf910937d4e60fdaf79a1
[ "MIT" ]
14
2020-01-13T21:20:07.000Z
2022-03-31T02:11:26.000Z
# Authors: Christian Thurau # License: BSD 3 Clause """ PyMF Non-negative Matrix Factorization. NMF: Class for Non-negative Matrix Factorization [1] Lee, D. D. and Seung, H. S. (1999), Learning the Parts of Objects by Non-negative Matrix Factorization, Nature 401(6755), 788-799. """ import numpy as np import logging import logging.config import scipy.sparse import scipy.optimize from cvxopt import solvers, base from base import PyMFBase from svd import pinv __all__ = ["NMF", "RNMF", "NMFALS", "NMFNNLS"] class NMF(PyMFBase): """ NMF(data, num_bases=4) Non-negative Matrix Factorization. Factorize a data matrix into two matrices s.t. F = | data - W*H | = | is minimal. H, and W are restricted to non-negative data. Uses the classicial multiplicative update rule. Parameters ---------- data : array_like, shape (_data_dimension, _num_samples) the input data num_bases: int, optional Number of bases to compute (column rank of W and row rank of H). 4 (default) Attributes ---------- W : "data_dimension x num_bases" matrix of basis vectors H : "num bases x num_samples" matrix of coefficients ferr : frobenius norm (after calling .factorize()) Example ------- Applying NMF to some rather stupid data set: >>> import numpy as np >>> data = np.array([[1.0, 0.0, 2.0], [0.0, 1.0, 1.0]]) >>> nmf_mdl = NMF(data, num_bases=2, niter=10) >>> nmf_mdl.factorize() The basis vectors are now stored in nmf_mdl.W, the coefficients in nmf_mdl.H. To compute coefficients for an existing set of basis vectors simply copy W to nmf_mdl.W, and set compute_w to False: >>> data = np.array([[1.5], [1.2]]) >>> W = np.array([[1.0, 0.0], [0.0, 1.0]]) >>> nmf_mdl = NMF(data, num_bases=2) >>> nmf_mdl.W = W >>> nmf_mdl.factorize(niter=20, compute_w=False) The result is a set of coefficients nmf_mdl.H, s.t. data = W * nmf_mdl.H. """ def _update_h(self): # pre init H1, and H2 (necessary for storing matrices on disk) H2 = np.dot(np.dot(self.W.T, self.W), self.H) + 10**-9 self.H *= np.dot(self.W.T, self.data[:,:]) self.H /= H2 def _update_w(self): # pre init W1, and W2 (necessary for storing matrices on disk) W2 = np.dot(np.dot(self.W, self.H), self.H.T) + 10**-9 self.W *= np.dot(self.data[:,:], self.H.T) self.W /= W2 self.W /= np.sqrt(np.sum(self.W**2.0, axis=0)) class RNMF(PyMFBase): """ RNMF(data, num_bases=4) Non-negative Matrix Factorization. Factorize a data matrix into two matrices s.t. F = | data - W*H | = | is minimal. H, and W are restricted to non-negative values. Uses the classicial multiplicative update rule. Parameters ---------- data : array_like, shape (_data_dimension, _num_samples) the input data num_bases: int, optional Number of bases to compute (column rank of W and row rank of H). 4 (default) Attributes ---------- W : "data_dimension x num_bases" matrix of basis vectors H : "num bases x num_samples" matrix of coefficients ferr : frobenius norm (after calling .factorize()) Example ------- Applying NMF to some rather stupid data set: >>> import numpy as np >>> data = np.array([[1.0, 0.0, 2.0], [0.0, 1.0, 1.0]]) >>> nmf_mdl = RNMF(data, num_bases=2) >>> nmf_mdl.factorize() The basis vectors are now stored in nmf_mdl.W, the coefficients in nmf_mdl.H. To compute coefficients for an existing set of basis vectors simply copy W to nmf_mdl.W, and set compute_w to False: >>> data = np.array([[1.5], [1.2]]) >>> W = np.array([[1.0, 0.0], [0.0, 1.0]]) >>> nmf_mdl = RNMF(data, num_bases=2) >>> nmf_mdl.W = W >>> nmf_mdl.factorize(niter=20, compute_w=False) The result is a set of coefficients nmf_mdl.H, s.t. data = W * nmf_mdl.H. """ def __init__(self, data, num_bases=4, lamb=2.0): # call inherited method PyMFBase.__init__(self, data, num_bases=num_bases) self._lamb = lamb def soft_thresholding(self, X, lamb): X = np.where(np.abs(X) <= lamb, 0.0, X) X = np.where(X > lamb, X - lamb, X) X = np.where(X < -1.0*lamb, X + lamb, X) return X def _init_h(self): self.H = np.random.random((self._num_bases, self._num_samples)) self.H[:,:] = 1.0 # normalized bases Wnorm = np.sqrt(np.sum(self.W**2.0, axis=0)) self.W /= Wnorm for i in range(self.H.shape[0]): self.H[i,:] *= Wnorm[i] self._update_s() def _update_s(self): self.S = self.data - np.dot(self.W, self.H) self.S = self.soft_thresholding(self.S, self._lamb) def _update_h(self): # pre init H1, and H2 (necessary for storing matrices on disk) H1 = np.dot(self.W.T, self.S - self.data) H1 = np.abs(H1) - H1 H1 /= (2.0* np.dot(self.W.T, np.dot(self.W, self.H))) self.H *= H1 # adapt S self._update_s() def _update_w(self): # pre init W1, and W2 (necessary for storing matrices on disk) W1 = np.dot(self.S - self.data, self.H.T) #W1 = np.dot(self.data - self.S, self.H.T) W1 = np.abs(W1) - W1 W1 /= (2.0 * (np.dot(self.W, np.dot(self.H, self.H.T)))) self.W *= W1 class NMFALS(PyMFBase): """ NMFALS(data, num_bases=4) Non-negative Matrix Factorization. Factorize a data matrix into two matrices s.t. F = | data - W*H | = | is minimal. H, and W are restricted to non-negative data. Uses the an alternating least squares procedure (quite slow for larger data sets) and cvxopt, similar to aa. Parameters ---------- data : array_like, shape (_data_dimension, _num_samples) the input data num_bases: int, optional Number of bases to compute (column rank of W and row rank of H). 4 (default) Attributes ---------- W : "data_dimension x num_bases" matrix of basis vectors H : "num bases x num_samples" matrix of coefficients ferr : frobenius norm (after calling .factorize()) Example ------- Applying NMF to some rather stupid data set: >>> import numpy as np >>> data = np.array([[1.0, 0.0, 2.0], [0.0, 1.0, 1.0]]) >>> nmf_mdl = NMFALS(data, num_bases=2) >>> nmf_mdl.factorize(niter=10) The basis vectors are now stored in nmf_mdl.W, the coefficients in nmf_mdl.H. To compute coefficients for an existing set of basis vectors simply copy W to nmf_mdl.W, and set compute_w to False: >>> data = np.array([[1.5], [1.2]]) >>> W = np.array([[1.0, 0.0], [0.0, 1.0]]) >>> nmf_mdl = NMFALS(data, num_bases=2) >>> nmf_mdl.W = W >>> nmf_mdl.factorize(niter=1, compute_w=False) The result is a set of coefficients nmf_mdl.H, s.t. data = W * nmf_mdl.H. """ def _update_h(self): def updatesingleH(i): # optimize alpha using qp solver from cvxopt FA = base.matrix(np.float64(np.dot(-self.W.T, self.data[:,i]))) al = solvers.qp(HA, FA, INQa, INQb) self.H[:,i] = np.array(al['x']).reshape((1,-1)) # float64 required for cvxopt HA = base.matrix(np.float64(np.dot(self.W.T, self.W))) INQa = base.matrix(-np.eye(self._num_bases)) INQb = base.matrix(0.0, (self._num_bases,1)) map(updatesingleH, xrange(self._num_samples)) def _update_w(self): def updatesingleW(i): # optimize alpha using qp solver from cvxopt FA = base.matrix(np.float64(np.dot(-self.H, self.data[i,:].T))) al = solvers.qp(HA, FA, INQa, INQb) self.W[i,:] = np.array(al['x']).reshape((1,-1)) # float64 required for cvxopt HA = base.matrix(np.float64(np.dot(self.H, self.H.T))) INQa = base.matrix(-np.eye(self._num_bases)) INQb = base.matrix(0.0, (self._num_bases,1)) map(updatesingleW, xrange(self._data_dimension)) self.W = self.W/np.sum(self.W, axis=1) class NMFNNLS(PyMFBase): """ NMFNNLS(data, num_bases=4) Non-negative Matrix Factorization. Factorize a data matrix into two matrices s.t. F = | data - W*H | = | is minimal. H, and W are restricted to non-negative data. Uses the Lawsons and Hanson's algorithm for non negative constrained least squares (-> also see scipy.optimize.nnls) Parameters ---------- data : array_like, shape (_data_dimension, _num_samples) the input data num_bases: int, optional Number of bases to compute (column rank of W and row rank of H). 4 (default) Attributes ---------- W : "data_dimension x num_bases" matrix of basis vectors H : "num bases x num_samples" matrix of coefficients ferr : frobenius norm (after calling .factorize()) Example ------- Applying NMF to some rather stupid data set: >>> import numpy as np >>> data = np.array([[1.0, 0.0, 2.0], [0.0, 1.0, 1.0]]) >>> nmf_mdl = NMFNNLS(data, num_bases=2) >>> nmf_mdl.factorize(niter=10) The basis vectors are now stored in nmf_mdl.W, the coefficients in nmf_mdl.H. To compute coefficients for an existing set of basis vectors simply copy W to nmf_mdl.W, and set compute_w to False: >>> data = np.array([[1.5], [1.2]]) >>> W = np.array([[1.0, 0.0], [0.0, 1.0]]) >>> nmf_mdl = NMFNNLS(data, num_bases=2) >>> nmf_mdl.W = W >>> nmf_mdl.factorize(niter=1, compute_w=False) The result is a set of coefficients nmf_mdl.H, s.t. data = W * nmf_mdl.H. """ def _update_h(self): def updatesingleH(i): self.H[:,i] = scipy.optimize.nnls(self.W, self.data[:,i])[0] map(updatesingleH, xrange(self._num_samples)) def _update_w(self): def updatesingleW(i): self.W[i,:] = scipy.optimize.nnls(self.H.T, self.data[i,:].T)[0] map(updatesingleW, xrange(self._data_dimension)) def _test(): import doctest doctest.testmod() if __name__ == "__main__": _test()
34.977918
86
0.554293
1,605
11,088
3.723988
0.122118
0.040154
0.010038
0.016731
0.80425
0.768446
0.7477
0.734817
0.71909
0.717417
0
0.028827
0.314845
11,088
316
87
35.088608
0.757931
0.553391
0
0.285714
0
0
0.007562
0
0
0
0
0
0
1
0.202381
false
0
0.107143
0
0.369048
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
0
0
0
0
0
7
69dbb7ccb70650348502e838d71fddbbb61d1bd0
159
py
Python
starlingx-dashboard/starlingx-dashboard/starlingx_dashboard/horizon/tables/__init__.py
MarioCarrilloA/gui
03a425f1bc1e4671a110f91eb88c272969dba57c
[ "Apache-2.0" ]
null
null
null
starlingx-dashboard/starlingx-dashboard/starlingx_dashboard/horizon/tables/__init__.py
MarioCarrilloA/gui
03a425f1bc1e4671a110f91eb88c272969dba57c
[ "Apache-2.0" ]
null
null
null
starlingx-dashboard/starlingx-dashboard/starlingx_dashboard/horizon/tables/__init__.py
MarioCarrilloA/gui
03a425f1bc1e4671a110f91eb88c272969dba57c
[ "Apache-2.0" ]
null
null
null
from starlingx_dashboard.horizon.tables.actions import FixedWithQueryFilter # noqa from starlingx_dashboard.horizon.tables.actions import LimitAction # noqa
53
83
0.861635
18
159
7.5
0.555556
0.192593
0.325926
0.42963
0.711111
0.711111
0.711111
0
0
0
0
0
0.08805
159
2
84
79.5
0.931034
0.056604
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
0
1
1
0
1
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
0
0
0
8
69ee0718f055cd622b37f02c5907911c4e1c5573
2,227
py
Python
parlai/agents/programr/mappings/storage/brainstorage.py
roholazandie/ParlAI
32352cab81ecb666aefd596232c5ed9f33cbaeb9
[ "MIT" ]
null
null
null
parlai/agents/programr/mappings/storage/brainstorage.py
roholazandie/ParlAI
32352cab81ecb666aefd596232c5ed9f33cbaeb9
[ "MIT" ]
null
null
null
parlai/agents/programr/mappings/storage/brainstorage.py
roholazandie/ParlAI
32352cab81ecb666aefd596232c5ed9f33cbaeb9
[ "MIT" ]
null
null
null
class BrainStorage: def __init__(self): pass def initialize(self): raise NotImplementedError("Should not call this base method, implementation missing") def save_properties(self): raise NotImplementedError("Should not call this base method, implementation missing") def load_properties(self, collection_name): raise NotImplementedError("Should not call this base method, implementation missing") def load_denormal(self, collection_name): raise NotImplementedError("Should not call this base method, implementation missing") def load_normal(self, collection_name): raise NotImplementedError("Should not call this base method, implementation missing") def load_gender(self, collection_name): raise NotImplementedError("Should not call this base method, implementation missing") def load_person(self, collection_name): raise NotImplementedError("Should not call this base method, implementation missing") def load_maps(self, collection_name): raise NotImplementedError("Should not call this base method, implementation missing") def load_variables(self, collection_name): raise NotImplementedError("Should not call this base method, implementation missing") def save_variable(self, collection_name, key, value): raise NotImplementedError("Should not call this base method, implementation missing") def save_gender(self, genders): raise NotImplementedError("Should not call this base method, implementation missing") def save_normal(self, normals): raise NotImplementedError("Should not call this base method, implementation missing") def save_denormal(self, denormals): raise NotImplementedError("Should not call this base method, implementation missing") def save_person(self, persons): raise NotImplementedError("Should not call this base method, implementation missing") def save_maps(self, maps): raise NotImplementedError("Should not call this base method, implementation missing") def save_variables(self, variables): raise NotImplementedError("Should not call this base method, implementation missing")
42.826923
93
0.746295
253
2,227
6.462451
0.13834
0.234862
0.293578
0.322936
0.850765
0.850765
0.850765
0.850765
0.850765
0.850765
0
0
0.187247
2,227
52
94
42.826923
0.903315
0
0
0.457143
0
0
0.402154
0
0
0
0
0
0
1
0.485714
false
0.028571
0
0
0.514286
0
0
0
0
null
1
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
0
0
0
0
0
10
384ee6ca346b4215dbad7143cfbf3acefa7f1de5
12,033
py
Python
tests/response_data.py
universalappfactory/pyhearthis
29c3a7e051d2937ad2fddb48d083e6a48b298014
[ "MIT" ]
null
null
null
tests/response_data.py
universalappfactory/pyhearthis
29c3a7e051d2937ad2fddb48d083e6a48b298014
[ "MIT" ]
null
null
null
tests/response_data.py
universalappfactory/pyhearthis
29c3a7e051d2937ad2fddb48d083e6a48b298014
[ "MIT" ]
null
null
null
WAVEFORM_RESPONSE_DATA = '''[187,191,215,197,211,223,229,236,229,219,188,169,176,167,166,230,240,218,197,216,209,227,229,237,238,245,193,201,208,255,235,242,240,230,212,221,200,207,223,235,235,234,226,188,193,210,229,240,234,222,225,210,202,204,233,237,243,225,220,213,211,215,229,240,241,230,222,209,196,227,220,231,162,161,155,164,188,174,180,169,190,214,216,225,212,177,202,234,227,230,211,216,198,197,196,209,226,231,229,225,206,192,196,210,210,217,230,198,183,173,215,214,199,205,207,207,227,213,207,206,224,200,206,212,223,217,197,187,184,229,212,189,200,214,209,210,215,208,200,219,222,224,187,197,193,180,174,149,158,174,178,187,177,187,191,226,224,203,215,225,216,240,211,219,226,215,212,238,194,215,215,190,193,216,218,235,215,220,225,211,236,226,220,228,243,234,225,219,222,235,217,219,216,224,235,203,213,199,211,206,215,206,213,247,221,218,235,234,216,210,211,238,200,201,217,210,210,195,199,221,198,240,214,223,234,224,235,221,217,201,201,234,210,221,204,197,204,219,212,224,226,213,207,192,246,219,226,210,203,212,203,226,210,208,212,206,213,221,209,209,227,216,225,227,237,212,210,214,228,207,223,216,250,217,199,207,220,228,225,218,204,223,196,208,242,216,212,208,213,233,212,166,168,175,179,169,181,181,177,190,190,203,182,190,192,201,201,207,208,216,217,207,213,212,208,205,204,222,219,216,210,227,239,226,225,219,232,226,214,230,223,244,231,246,234,247,233,244,249,234,226,247,238,242,235,239,210,225,230,246,229,233,247,247,230,243,212,204,192,180,208,182,181,191,167,190,171,186,177,177,185,167,219,240,238,242,230,242,220,227,239,214,245,238,225,241,255,255,225,225,240,242,211,237,234,236,230,237,217,224,234,247,202,217,231,237,228,233,194,200,200,208,204,215,213,212,211,216,209,220,214,207,203,209,217,219,227,227,229,237,228,147,150,147,153,221,232,239,236,235,227,222,225,222,220,222,215,231,228,229,230,231,224,221,234,228,234,229,228,219,200,216,216,214,224,229,235,231,233,238,231,230,228,235,230,222,227,218,194,215,227,239,239,221,226,236,220,227,228,231,228,222,212,201,211,216,204,223,242,242,243,236,218,236,208,229,212,220,206,204,218,224,217,198,212,213,220,250,243,228,227,246,242,249,224,237,241,231,239,216,240,221,214,228,209,237,232,249,232,222,219,215,216,213,244,209,222,239,215,234,220,222,214,210,225,235,244,219,244,233,237,235,248,218,229,236,247,219,243,230,247,241,237,237,236,235,240,241,234,226,242,242,242,237,242,240,247,251,247,248,238,224,242,228,243,212,221,224,226,248,219,211,192,186,200,210,205,192,216,223,218,222,221,249,255,244,233,227,235,249,226,219,220,235,236,230,226,230,234,241,240,244,238,244,225,243,234,244,226,242,228,221,191,181,175,168,163,184,171,183,208,231,232,239,249,226,233,232,224,225,237,246,226,232,238,218,222,226,239,255,248,239,242,244,232,242,229,225,234,225,220,234,218,239,236,235,226,230,214,204,223,202,199,218,208,229,208,227,242,222,237,219,223,217,213,205,222,242,249,224,230,238,238,234,237,221,242,231,206,238,228,235,206,240,217,226,236,208,210,231,244,237,250,231,234,249,251,244,227,238,226,229,212,219,214,228,221,215,208,212,227,217,221,225,219,228,210,237,217,221,197,221,215,219,213,210,223,236,208,222,240,220,233,217,209,226,231,212,232,231,233,210,198,226,225,223,215,217,232,229,229,213,218,234,218,230,240,223,220,224,229,251,221,219,232,229,214,212,234,220,208,211,234,236,217,218,161,163,158,168,227,231,248,225,220,243,236,232,235,243,238,236,222,228,230,229,230,212,235,232,237,219,222,236,223,226,216,234,239,230,224,226,214,210,220,239,229,215,216,224,240,209,226,231,233,214,211,226,230,211,206,223,233,237,200,201,220,216,198,191,203,202,193,236,216,208,207,211,219,210,219,197,224,210,192,226,202,204,205,212,196,206,214,215,188,195,204,202,208,183,199,192,197,195,201,193,224,188,167,176,187,179,192,196,218,207,191,202,221,213,223,211,208,206,212,198,196,208,211,201,192,197,205,203,195,183,207,198,203,186,188,205,205,195,193,216,197,189,189,186,191,188,194,192,198,197,201,186,206,197,203,220,213,213,195,209,212,202,213,210,211,227,184,163,170,157,170,163,170,177,152,167,160,187,168,170,171,180,172,179,187,171,184,215,206,218,216,201,240,215,201,223,202,194,209,206,213,235,208,226,218,210,220,209,214,196,216,223,210,215,190,218,201,216,210,205,210,221,205,216,210,200,209,228,184,215,235,232,227,227,223,227,223,218,231,224,228,222,222,235,229,238,226,227,225,229,236,229,229,241,239,238,240,232,248,232,239,244,235,235,246,242,238,226,233,239,238,230,222,230,225,234,232,231,235,235,242,240,230,226,228,233,230,239,240,244,230,228,245,242,232,233,229,242,237,229,229,245,242,233,233,237,230,231,228,237,242,236,242,236,238,233,240,237,228,232,243,241,238,240,239,232,233,233,237,239,246,226,233,240,238,235,240,239,244,233,237,232,236,248,241,234,232,242,239,239,206,188,192,193,195,193,193,201,206,215,228,228,219,243,229,240,177,212,212,214,214,224,210,218,199,245,214,216,224,219,225,220,229,214,220,234,218,213,218,222,227,234,227,217,224,216,180,163,210,220,208,217,227,226,212,218,220,218,208,223,209,226,210,211,213,217,212,179,171,186,173,188,207,208,204,211,192,200,200,200,207,200,200,211,219,223,241,231,198,221,219,222,195,170,213,226,229,197,199,214,219,224,227,197,197,209,200,212,161,159,202,222,224,223,239,214,200,214,217,212,208,181,166,169,180,216,209,216,221,224,213,218,203,195,230,203,210,195,204,201,207,215,211,203,200,189,200,207,216,207,204,210,201,194,210,219,210,210,190,205,197,189,188,206,184,181,182,191,213,198,204,203,198,203,200,202,190,218,221,193,189,207,207,196,193,208,190,189,226,205,212,200,198,211,198,196,192,208,195,184,187,192,162,159,171,157,168,163,162,153,175,195,216,207,218,210,213,208,228,211,222,202,228,215,209,211,226,224,219,217,224,209,223,212,220,218,221,203,208,224,219,211,204,204,214,221,206,212,200,215,227,214,199,206,204,207,216,209,213,211,215,211,227,218,198,172,158,175,180,181,211,206,181,201,198,224,226,228,220,220,225,224,230,230,215,230,207,214,212,211,215,215,218,233,214,225,207,199,215,220,222,211,209,229,217,214,221,213,189,217,216,215,235,216,218,202,160,169,177,176,162,190,184,171,185,196,195,180,191,229,202,229,221,228,203,213,210,201,211,213,217,215,222,237,237,196,199,209,213,233,229,217,227,222,217,229,228,208,209,219,191,204,211,222,217,192,205,206,198,200,209,212,207,204,210,221,185,200,218,205,209,206,230,200,200,213,196,198,208,198,205,204,172,168,171,177,181,204,205,207,201,201,211,211,206,201,159,213,202,203,202,209,221,166,161,153,159,155,174,162,168,168,179,206,201,201,201,194,210,206,211,190,201,201,192,221,211,212,223,216,216,231,213,174,167,173,167,165,173,172,164,170,163,163,171,177,167,178,172,161,179,190,191,232,212,221,224,220,216,215,241,227,235,208,217,235,208,218,224,215,221,232,220,220,194,202,185,188,215,220,210,212,208,239,226,219,228,221,242,212,230,216,223,227,169,175,216,208,223,233,208,213,228,214,211,209,224,237,210,225,219,226,235,216,227,206,193,175,153,148,167,157,154,215,216,221,229,211,203,215,193,217,195,205,187,224,200,178,196,202,201,226,201,181,203,207,187,175,161,168,168,175,200,235,224,227,220,218,227,207,206,199,209,204,200,211,172,172,162,161,201,202,197,202,194,209,202,233,203,214,209,215,227,198,207,206,180,173,180,178,195,214,212,202,198,201,209,196,205,210,190,206,206,205,229,196,211,209,211,214,202,191,181,185,179,179,189,189,192,199,215,194,207,199,203,202,196,152,151,154,156,160,159,174,165,180,207,199,210,217,210,204,201,204,206,215,200,207,210,223,199,218,229,225,214,237,182,170,171,163,166,157,154,164,162,160,178,169,202,172,167,171,179,175,190,177,185,222,217,199,208,223,217,210,231,220,215,209,221,207,218,221,219,209,214,208,198,212,218,207,212,225,226,209,213,226,230,178,193,197,194,208,188,179,174,177,174,186,182,210,198,217,209,194,178,171,187,178,191,196,197,187,192,205,195,186,174,190,193,178,185,186,195,193,205,189,184,169,179,188,187,207,192,189,204,199,196,177,196,186,178,185,189,188,188,205,210,194,186,177,184,177,183,213,206,207,187,198,192,182,202,202,220,219,201,189,209,198,200,191,186,185,197,195,201,203,195,201,209,199,201,201,187,196,179,188,190,188,205,185,212,191,188,181,184,189,181,190,201,210,219,209,199,182,182,190,217,187,196,196,179,193,211,202,181,191,203,186,194,189,209,198,223,200,204,228,189,215,203,234,210,227,216,220,211,216,217,244,230,223,225,231,243,244,238,231,217,215,237,223,232,220,224,243,217,216,234,233,200,220,228,217,238,226,215,223,227,220,185,182,174,187,223,222,235,235,245,231,201,229,239,222,215,240,223,214,222,214,221,213,221,237,225,222,207,223,230,220,229,222,243,219,216,220,228,223,217,211,232,212,239,217,234,214,222,222,229,214,221,214,215,233,238,236,209,217,217,213,224,224,226,211,221,228,211,213,218,235,210,160,165,162,176,165,211,213,224,237,215,188,182,203,199,200,209,201,214,215,233,205,194,217,251,227,223,214,220,214,214,215,204,209,215,206,200,217,211,203,202,199,198,207,209,213,210,217,240,201,195,197,190,214,210,212,205,208,201,197,203,218,224,241,214,200,205,199,207,198,154,142,156,165,179,163,162,218,207,205,203,200,202,204,200,212,206,201,208,201,202,193,203,200,207,194,194,210,196,204,195,198,208,208,197,202,214,207,203,197,209,203,223,206,193,195,204,191,211,209,214,203,201,213,210,202,215,201,206,195,208,174,173,187,194,185,179,191,199,200,201,207,204,203,193,187,198,202,201,204,176,200,195,193,201,210,206,213,223,218,213,210,201,203,213,211,217,215,221,209,210,187,186,201,197,200,200,164,161,169,165,156,181,180,183,168,172,193,171,193,186,182,198,183,181,176,174,205,200,195,206,187,218,186,192,187,198,217,236,220,227,220,222,218,234,204,173,209,215,213,214,217,217,217,206,220,209,216,221,217,209,195,208,220,215,214,226,227,194,211,179,196,165,176,188,201,199,182,183,190,193,199,203,190,213,207,212,218,214,213,213,205,211,205,160,158,156,173,191,172,178,179,177,184,181,188,187,182,204,241,211,211,212,208,210,197,216,228,203,207,212,206,199,213,202,221,200,206,208,208,213,215,218,234,180,188,191,171,174,178,172,186,181,181,213,216,213,212,210,211,216,204,219,212,217,207,213,196,203,208,219,212,211,218,213,211,212,217,215,214,206,217,219,221,216,224,218,224,232,227,228,232,219,214,227,234,232,234,229,235,227,234,227,237,221,223,230,225,232,247,235,229,230,224,234,232,242,228,227,216,226,226,229,237,228,223,221,229,221,228,225,253,227,232,230,238,236,226,235,230,236,235,229,234,225,229,228,236,246,222,232,222,230,229,230,226,227,231,231,224,220,229,236,234,229,226,226,210,220,215,205,219,218,230,227,219,221,215,224,202,207,207,220,201,216,205,218,211,196,226,230,227,227,235,240,229,208,226,238,252,225,235,244,242,217,227,239,219,210,230,191,191,210,184,183,210,217,229,227,217,205,222,223,208,221,214,229,222,204,221,233,233,217,217,234,211,209,227,222,218,225,219,228,207,188,192,201,181,185,195,192,187,194,218,227,232,223,221,221,228,208,201,224,231,230,208,229,227,224,202,228,230,226,198,226,228,222,208,209,237,229,214,228,231,227,213,204,225,230,186,188,202,195,193,197,202,195,212,217,217,222,225,215,220,209,204,213,218,207,168,183,196,190,240,209,209,210,202,214,207,229,212,206,216,219,211,218,218,219,217,227,221,215,202,209,215,220,215,230,214,221,203,204,210,208,227,208,225,213,196,177,181,198,208,182,190,186,189,198,192,200,202,198,186,192,162,157,165,165,206,211,202,210,214,206,205,226,223,209,213,201,195,204,213,237,208,221,229,191,190,216,212,226,213,217,220,211,208,205,209,199,212,206,216,213,216,212,203,193,208,209,213,215,188,182,203,206,206,219,210,199,202,205,208,203,197,204,203,209,213,190,192,195,220,194,200,215,218,219,199,214,204,196,213,212,213,190,198,208,202,207,205,200,226,212,210,213,207,226,209,202,203,200,214,200,199,237,222,227,218,217,194,211,212,228,233,236,225,227,225,207,209,219,235,234,237,222,219,204,220,236,236,233,229,226,229,219,207,226,218,223,225,223,229,207,200,226,222,219,221,230,232,222,217,231,229,229,230,229,221,231,202,209,216,223,224,230,217,230,213,207,229,224,211,220,229,223,224,194,222,220,231,220,229,228,211,214,211,211,221,228,227,229,208,199,216,230,220,232,218,222,224,213,212,216,223,231,237,220,212,205,207,218,213,220,227,227,232,225,196,212,232,228,216,222,227,217,204,216,213,224,230,235,225,225,215,197,233,207,222,226,235,238,224,185,128]'''
6,016.5
12,032
0.749771
3,003
12,033
3.003663
0.037629
0.004656
0.002993
0.002661
0
0
0
0
0
0
0
0.74813
0.000249
12,033
2
12,032
6,016.5
0.001663
0
0
0
0
1
0.997341
0.997341
0
0
0
0
0
1
0
false
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
1
0
0
1
1
1
1
0
0
0
0
1
1
1
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
3861ae18b9f3d8eb13bd13c983b8a0afcb588621
183
py
Python
summarize/data/paragraph_tokenizers/__init__.py
danieldeutsch/summarize
f36a86d58f381ff1f607f356dad3d6ef7b0e0224
[ "Apache-2.0" ]
15
2019-11-01T11:49:44.000Z
2021-01-19T06:59:32.000Z
summarize/data/paragraph_tokenizers/__init__.py
CogComp/summary-cloze
b38e3e8c7755903477fd92a4cff27125cbf5553d
[ "Apache-2.0" ]
2
2020-03-30T07:54:01.000Z
2021-11-15T16:27:42.000Z
summarize/data/paragraph_tokenizers/__init__.py
CogComp/summary-cloze
b38e3e8c7755903477fd92a4cff27125cbf5553d
[ "Apache-2.0" ]
3
2019-12-06T05:57:51.000Z
2019-12-11T11:34:21.000Z
from summarize.data.paragraph_tokenizers.paragraph_tokenizer import ParagraphTokenizer from summarize.data.paragraph_tokenizers.paragraph_word_tokenizer import ParagraphWordTokenizer
61
95
0.923497
19
183
8.631579
0.526316
0.158537
0.207317
0.317073
0.54878
0.54878
0
0
0
0
0
0
0.043716
183
2
96
91.5
0.937143
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
3888f75f351e24884fda5e20895149316b64e35d
32,931
py
Python
RouterConfiguration/Juniper/juniper_config_printer.py
nsg-ethz/Metha
929839994043d80e20151c1bed33f8a1e3c1f8bd
[ "MIT" ]
9
2021-02-24T12:17:28.000Z
2021-11-29T02:24:58.000Z
RouterConfiguration/Juniper/juniper_config_printer.py
nsg-ethz/Metha
929839994043d80e20151c1bed33f8a1e3c1f8bd
[ "MIT" ]
null
null
null
RouterConfiguration/Juniper/juniper_config_printer.py
nsg-ethz/Metha
929839994043d80e20151c1bed33f8a1e3c1f8bd
[ "MIT" ]
2
2021-03-08T22:21:50.000Z
2021-04-20T06:59:40.000Z
import os from utils import * from RouterConfiguration.Juniper.juniper_config_features import * from network_features import * feature_config = { ProtocolIndependentFeatures.STATIC_ROUTE: lambda network, interface: f'set routing-options static route {int_to_ip(network.address)}/{network.prefix} next-hop {interface.name}', OSPFFeatures.AREA_RANGE: lambda area, network, override_metric, restrict: f'set protocols ospf area {area} area-range {int_to_ip(network.address)}/{network.prefix} {override_metric} {restrict}', OSPFFeatures.AREA_LABEL_SWITCHED_PATH: lambda area, path, metric: f'set protocols ospf area {area} label-switched-path {path} metric {metric}', OSPFFeatures.NSSA_DEFAULT_LSA: lambda area, metric, metric_type, type_7: f'set protocols ospf area {area} nssa default-lsa default-metric {metric} {metric_type} {type_7}', OSPFFeatures.NSSA_NO_SUMMARIES: lambda area: f'set protocols ospf area {area} nssa no-summaries', OSPFFeatures.STUB_DEFAULT_METRIC: lambda area, metric: f'set protocols ospf area {area} stub default-metric {metric}', OSPFFeatures.STUB_NO_SUMMARIES: lambda area: f'set protocols ospf area {area} stub no-summaries', OSPFFeatures.EXTERNAL_PREFERENCE: lambda preference: f'set protocols ospf external-preference {preference}', OSPFFeatures.NO_RFC_1583: lambda: f'set protocols ospf no-rfc-1583', OSPFFeatures.REFERENCE_BANDWIDTH: lambda bandwidth: f'set protocols ospf reference-bandwidth {bandwidth}', OSPFFeatures.INTERFACE_LDP_SYNCHRONIZATION: lambda interface, disable: f'set protocols ospf area {interface.area} interface {interface.name} ldp-sychronization {disable}', OSPFFeatures.INTERFACE_LINK_PROTECTION: lambda interface: f'set protocols ospf area {interface.area} interface {interface.name} link-protection', OSPFFeatures.INTERFACE_METRIC: lambda interface, metric: f'set protocols ospf area {interface.area} interface {interface.name} metric {metric}', OSPFFeatures.INTERFACE_PASSIVE: lambda interface: f'set protocols ospf area {interface.area} interface {interface.name} passive', OSPFFeatures.INTERFACE_PRIORITY: lambda interface, priority: f'set protocols ospf area {interface.area} interface {interface.name} priority {priority}', OSPFFeatures.INTERFACE_TE_METRIC: lambda interface, metric: f'set protocols ospf area {interface.area} interface {interface.name} te-metric {metric}', OSPFFeatures.REDISTRIBUTE_DIRECT: lambda: f'set protocols ospf export send-direct', OSPFFeatures.REDISTRIBUTE_STATIC: lambda: f'set protocols ospf export send-static', OSPFFeatures.REDISTRIBUTE_BGP: lambda: f'set protocols ospf export send-bgp', OSPFFeatures.EXPORT: lambda: f'set protocols ospf export export-ospf', OSPFFeatures.IMPORT: lambda: f'set protocols ospf import import-ospf', BGPFeatures.ACCEPTED_PREFIX_LIMIT: lambda limit: f'set protocols bgp family inet any accepted-prefix-limit maximum {limit}', BGPFeatures.ADVERTISE_EXTERNAL: lambda group: f'set protocols bgp group {group} advertise-external', BGPFeatures.ADVERTISE_INACTIVE: lambda: f'set protocols bgp advertise-inactive', BGPFeatures.ADVERTISE_PEER_AS: lambda: f'set protocols bgp advertise-peer-as', BGPFeatures.AS_OVERRIDE: lambda group: f'set protocols bgp group {group} as-override', BGPFeatures.CLUSTER: lambda: f'', BGPFeatures.DAMPING: lambda: f'set protocols bgp damping', BGPFeatures.ENFORCE_FIRST_AS: lambda: f'set protocols bgp group EBGP enforce-first-as', BGPFeatures.LOCAL_AS: lambda AS, option: f'set protocols bgp local-as {AS.num} {option}', BGPFeatures.METRIC_OUT: lambda metric: f'set protocols bgp metric-out {metric}', BGPFeatures.MULTIHOP: lambda: f'', BGPFeatures.NO_CLIENT_REFLECT: lambda: f'set protocols bgp no-client-reflect', BGPFeatures.PASSIVE: lambda: f'set protocols bgp passive', BGPFeatures.PATH_SELECTION: lambda option: f'set protocols bgp path-selection {option}', BGPFeatures.REMOVE_PRIVATE: lambda: f'set protocols bgp remove-private', BGPFeatures.TCP_MSS: lambda size: f'set protocols bgp tcp-mss {size}', BGPFeatures.ADD_PATH: lambda group, options: f'set protocols bgp group {group} family inet unicast add-path {options}', BGPFeatures.LOOPS: lambda loops: f'set protocols bgp family inet unicast loops {loops}', BGPFeatures.PREFIX_LIMIT: lambda limit: f'set protocols bgp family inet any prefix-limit maximum {limit}', BGPFeatures.REDISTRIBUTE_DIRECT: lambda: f'set protocols bgp export send-direct', BGPFeatures.REDISTRIBUTE_STATIC: lambda: f'set protocols bgp export send-static', BGPFeatures.REDISTRIBUTE_OSPF: lambda: f'set protocols bgp export send-ospf', BGPFeatures.IMPORT: lambda: f'set protocols bgp import import-bgp', BGPFeatures.EXPORT: lambda: f'set protocols bgp export export-bgp', BGPFeatures.LOCAL_PREFERENCE: lambda pref: f'set protocols bgp local-preference {pref}', BGPFeatures.NEIGHBOUR_POLICY_EXPORT: lambda neighbour, policy: f'set protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} export {policy.name}', BGPFeatures.NEIGHBOUR_POLICY_IMPORT: lambda neighbour, policy: f'set protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} import {policy.name}', PolicyFeatures.FROM_AREA: lambda policy, term, area: f'set policy-options policy-statement {policy.name} term {term} from area {area}', PolicyFeatures.FROM_AS_PATH: lambda policy, term, as_path: f'set policy-options policy-statement {policy.name} term {term} from as-path {as_path}', PolicyFeatures.FROM_AS_PATH_GROUP: lambda policy, term, group: f'set policy-options policy-statement {policy.name} term {term} from as-path-group {group}', PolicyFeatures.FROM_COLOR: lambda policy, term, color: f'set policy-options policy-statement {policy.name} term {term} from color {color}', PolicyFeatures.FROM_COMMUNITY: lambda policy, term, community: f'set policy-options policy-statement {policy.name} term {term} from community {community}', PolicyFeatures.FROM_FAMILY: lambda policy, term, family: f'set policy-options policy-statement {policy.name} term {term} from family {family}', PolicyFeatures.FROM_INSTANCE: lambda: f'', PolicyFeatures.FROM_INTERFACE: lambda policy, term, interface: f'set policy-options policy-statement {policy.name} term {term} from interface {interface.name}', PolicyFeatures.FROM_LEVEL: lambda: f'', PolicyFeatures.FROM_LOCAL_PREFERENCE: lambda policy, term, preference: f'set policy-options policy-statement {policy.name} term {term} from local-preference {preference}', PolicyFeatures.FROM_METRIC: lambda policy, term, metric: f'set policy-options policy-statement {policy.name} term {term} from metric {metric}', PolicyFeatures.FROM_NEIGHBOUR: lambda policy, term, neighbour: f'set policy-options policy-statement {policy.name} term {term} from neighbor {int_to_ip(neighbour.address)}', PolicyFeatures.FROM_ORIGIN: lambda policy, term, origin: f'set policy-options policy-statement {policy.name} term {term} from origin {origin}', PolicyFeatures.FROM_POLICY: lambda policy, term, policy2: f'set policy-options policy-statement {policy.name} term {term} from policy {policy2}', PolicyFeatures.FROM_PREFIX_LIST: lambda policy, term, prefix_list: f'set policy-options policy-statement {policy.name} term {term} from prefix-list {prefix_list}', PolicyFeatures.FROM_PREFIX_LIST_FILTER: lambda policy, term, prefix_list, match_type: f'set policy-options policy-statement {policy.name} term {term} from prefix-list-filter {prefix_list} {match_type}', PolicyFeatures.FROM_PROTOCOL: lambda policy, term, protocol: f'set policy-options policy-statement {policy.name} term {term} from protocol {protocol}', PolicyFeatures.FROM_RIB: lambda policy, term, rib: f'set policy-options policy-statement {policy.name} term {term} from rib {rib}', PolicyFeatures.FROM_ROUTE_FILTER: lambda policy, term, net, match_type: f'set policy-options policy-statement {policy.name} term {term} from route-filter {int_to_ip(net.address)}/{net.prefix} {match_type}', PolicyFeatures.FROM_ROUTE_TYPE: lambda policy, term, route_type: f'set policy-options policy-statement {policy.name} term {term} from route-type {route_type}', PolicyFeatures.FROM_SOURCE_ADDRESS_FILTER: lambda policy, term, source_filter, match_type: f'set policy-options policy-statement {policy.name} term {term} from source-address-filter {source_filter} {match_type}', PolicyFeatures.FROM_TAG: lambda policy, term, tag: f'set policy-options policy-statement {policy.name} term {term} from tag {tag}', PolicyFeatures.FROM_NEXT_HOP: lambda policy, term, next_hop: f'set policy-options policy-statement {policy.name} term {term} from next-hop {int_to_ip(next_hop)}', PolicyFeatures.TO_LEVEL: lambda: f'', PolicyFeatures.TO_RIB: lambda: f'', PolicyFeatures.THEN_ACCEPT: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then accept', PolicyFeatures.THEN_AS_PATH_EXPAND: lambda policy, term, n: f'set policy-options policy-statement {policy.name} term {term} then as-path-expand last-as count {n}', PolicyFeatures.THEN_AS_PATH_PREPEND: lambda policy, term, as_path: f'set policy-options policy-statement {policy.name} term {term} then as-path-prepend {as_path}', PolicyFeatures.THEN_COLOR: lambda policy, term, add, color: f'set policy-options policy-statement {policy.name} term {term} then color {add} {color}', PolicyFeatures.THEN_COLOR2: lambda policy, term, add, color: f'set policy-options policy-statement {policy.name} term {term} then color2 {add} {color}', PolicyFeatures.THEN_COMMUNITY_ADD: lambda policy, term, community: f'set policy-options policy-statement {policy.name} term {term} then community add {community}', PolicyFeatures.THEN_COMMUNITY_DELETE: lambda policy, term, community: f'set policy-options policy-statement {policy.name} term {term} then community delete {community}', PolicyFeatures.THEN_COMMUNITY_SET: lambda policy, term, community: f'set policy-options policy-statement {policy.name} term {term} then community set {community}', PolicyFeatures.THEN_COS_NEXT_HOP_MAP: lambda policy, term, cos_map: f'set policy-options policy-statement {policy.name} term {term} then cos-next-hop-map {cos_map}', PolicyFeatures.THEN_DEFAULT_ACTION_ACCEPT: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then default-action accept', PolicyFeatures.THEN_DEFAULT_ACTION_REJECT: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then default-action reject', PolicyFeatures.THEN_EXTERNAL: lambda policy, term, metric_type: f'set policy-options policy-statement {policy.name} term {term} then external type {metric_type}', PolicyFeatures.THEN_FORWARDING_CLASS: lambda policy, term, fwd_class: f'set policy-options policy-statement {policy.name} term {term} then forwarding-class {fwd_class}', PolicyFeatures.THEN_INSTALL_NEXTHOP: lambda: f'', PolicyFeatures.THEN_LOCAL_PREFERENCE: lambda policy, term, pref: f'set policy-options policy-statement {policy.name} term {term} then local-preference {pref}', PolicyFeatures.THEN_METRIC: lambda policy, term, metric: f'set policy-options policy-statement {policy.name} term {term} then metric {metric}', PolicyFeatures.THEN_METRIC_ADD: lambda policy, term, metric: f'set policy-options policy-statement {policy.name} term {term} then metric add {metric}', PolicyFeatures.THEN_METRIC_EXPRESSION: lambda: f'', PolicyFeatures.THEN_METRIC_IGP: lambda policy, term, offset: f'set policy-options policy-statement {policy.name} term {term} then metric igp {offset}', PolicyFeatures.THEN_METRIC2: lambda policy, term, metric: f'set policy-options policy-statement {policy.name} term {term} then metric2 {metric}', PolicyFeatures.THEN_METRIC2_EXPRESSION: lambda: f'', PolicyFeatures.THEN_NEXT_HOP: lambda policy, term, next_hop: f'set policy-options policy-statement {policy.name} term {term} then next-hop {int_to_ip(next_hop)}', PolicyFeatures.THEN_NEXT_HOP_SELF: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then next-hop self', PolicyFeatures.THEN_NEXT_POLICY: lambda: f'', PolicyFeatures.THEN_NEXT_TERM: lambda: f'', PolicyFeatures.THEN_ORIGIN: lambda policy, term, origin: f'set policy-options policy-statement {policy.name} term {term} then origin {origin}', PolicyFeatures.THEN_PREFERENCE: lambda policy, term, pref: f'set policy-options policy-statement {policy.name} term {term} then preference {pref}', PolicyFeatures.THEN_PRIORITY: lambda policy, term, priority: f'set policy-options policy-statement {policy.name} term {term} then priority {priority}', PolicyFeatures.THEN_REJECT: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then reject', PolicyFeatures.THEN_TAG: lambda policy, term, tag: f'set policy-options policy-statement {policy.name} term {term} then tag {tag}', PolicyFeatures.POLICY_MATCH_FEATURE_BGP_OUT: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args), PolicyFeatures.POLICY_MATCH_FEATURE_BGP_IN: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args), PolicyFeatures.POLICY_MATCH_FEATURE_OSPF_OUT: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args), PolicyFeatures.POLICY_MATCH_FEATURE_OSPF_IN: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args), PolicyFeatures.POLICY_SET_FEATURE_BGP_OUT: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args), PolicyFeatures.POLICY_SET_FEATURE_BGP_IN: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args), PolicyFeatures.POLICY_SET_FEATURE_OSPF_OUT: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args), PolicyFeatures.POLICY_SET_FEATURE_OSPF_IN: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args) } feature_disable = { ProtocolIndependentFeatures.STATIC_ROUTE: lambda network, interface: f'delete routing-options static route {int_to_ip(network.address)}/{network.prefix} next-hop {interface.name}', OSPFFeatures.AREA_RANGE: lambda area, network, override_metric, restrict: f'delete protocols ospf area {area} area-range {int_to_ip(network.address)}/{network.prefix}', OSPFFeatures.AREA_LABEL_SWITCHED_PATH: lambda area, path, metric: f'delete protocols ospf area {area} label-switched-path {path} metric {metric}', OSPFFeatures.NSSA_DEFAULT_LSA: lambda area, metric, metric_type, type_7: f'delete protocols ospf area {area} nssa default-lsa', OSPFFeatures.NSSA_NO_SUMMARIES: lambda area: f'delete protocols ospf area {area} nssa no-summaries', OSPFFeatures.STUB_DEFAULT_METRIC: lambda area, metric: f'delete protocols ospf area {area} stub default-metric {metric}', OSPFFeatures.STUB_NO_SUMMARIES: lambda area: f'delete protocols ospf area {area} stub no-summaries', OSPFFeatures.EXTERNAL_PREFERENCE: lambda preference: f'delete protocols ospf external-preference {preference}', OSPFFeatures.NO_RFC_1583: lambda: f'delete protocols ospf no-rfc-1583', OSPFFeatures.REFERENCE_BANDWIDTH: lambda bandwidth: f'delete protocols ospf reference-bandwidth {bandwidth}', OSPFFeatures.INTERFACE_LDP_SYNCHRONIZATION: lambda interface, disable: f'delete protocols ospf area {interface.area} interface {interface.name} ldp-sychronization', OSPFFeatures.INTERFACE_LINK_PROTECTION: lambda interface: f'delete protocols ospf area {interface.area} interface {interface.name} link-protection', OSPFFeatures.INTERFACE_METRIC: lambda interface, metric: f'delete protocols ospf area {interface.area} interface {interface.name} metric {metric}', OSPFFeatures.INTERFACE_PASSIVE: lambda interface: f'delete protocols ospf area {interface.area} interface {interface.name} passive', OSPFFeatures.INTERFACE_PRIORITY: lambda interface, priority: f'delete protocols ospf area {interface.area} interface {interface.name} priority {priority}', OSPFFeatures.INTERFACE_TE_METRIC: lambda interface, metric: f'delete protocols ospf area {interface.area} interface {interface.name} te-metric {metric}', OSPFFeatures.REDISTRIBUTE_DIRECT: lambda: f'delete protocols ospf export send-direct', OSPFFeatures.REDISTRIBUTE_STATIC: lambda: f'delete protocols ospf export send-static', OSPFFeatures.REDISTRIBUTE_BGP: lambda: f'delete protocols ospf export send-bgp', OSPFFeatures.EXPORT: lambda: f'delete protocols ospf export export-ospf', OSPFFeatures.IMPORT: lambda: f'delete protocols ospf import import-ospf', BGPFeatures.ACCEPTED_PREFIX_LIMIT: lambda limit: f'delete protocols bgp family inet any accepted-prefix-limit maximum {limit}', BGPFeatures.ADVERTISE_EXTERNAL: lambda group: f'delete protocols bgp group {group} advertise-external', BGPFeatures.ADVERTISE_INACTIVE: lambda: f'delete protocols bgp advertise-inactive', BGPFeatures.ADVERTISE_PEER_AS: lambda: f'delete protocols bgp advertise-peer-as', BGPFeatures.AS_OVERRIDE: lambda group: f'delete protocols bgp group {group} as-override', BGPFeatures.CLUSTER: lambda: f'', BGPFeatures.DAMPING: lambda: f'delete protocols bgp damping', BGPFeatures.ENFORCE_FIRST_AS: lambda: f'delete protocols bgp group EBGP enforce-first-as', BGPFeatures.LOCAL_AS: lambda AS, option: f'delete protocols bgp local-as {AS.num}', BGPFeatures.METRIC_OUT: lambda metric: f'delete protocols bgp metric-out {metric}', BGPFeatures.MULTIHOP: lambda: f'', BGPFeatures.NO_CLIENT_REFLECT: lambda: f'delete protocols bgp no-client-reflect', BGPFeatures.PASSIVE: lambda: f'delete protocols bgp passive', BGPFeatures.PATH_SELECTION: lambda option: f'delete protocols bgp path-selection', BGPFeatures.REMOVE_PRIVATE: lambda: f'delete protocols bgp remove-private', BGPFeatures.TCP_MSS: lambda size: f'delete protocols bgp tcp-mss {size}', BGPFeatures.ADD_PATH: lambda group, options: f'delete protocols bgp group {group} family inet unicast add-path', BGPFeatures.LOOPS: lambda loops: f'delete protocols bgp family inet unicast loops {loops}', BGPFeatures.PREFIX_LIMIT: lambda limit: f'delete protocols bgp family inet any prefix-limit maximum {limit}', BGPFeatures.REDISTRIBUTE_DIRECT: lambda: f'delete protocols bgp export send-direct', BGPFeatures.REDISTRIBUTE_STATIC: lambda: f'delete protocols bgp export send-static', BGPFeatures.REDISTRIBUTE_OSPF: lambda: f'delete protocols bgp export send-ospf', BGPFeatures.IMPORT: lambda: f'delete protocols bgp import import-bgp', BGPFeatures.EXPORT: lambda: f'delete protocols bgp export export-bgp', BGPFeatures.LOCAL_PREFERENCE: lambda pref: f'delete protocols bgp local-preference {pref}', BGPFeatures.NEIGHBOUR_POLICY_EXPORT: lambda neighbour, policy: f'delete protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} export {policy.name}', BGPFeatures.NEIGHBOUR_POLICY_IMPORT: lambda neighbour, policy: f'delete protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} import {policy.name}', PolicyFeatures.FROM_AREA: lambda policy, term, area: f'delete policy-options policy-statement {policy.name} term {term} from area {area}', PolicyFeatures.FROM_AS_PATH: lambda policy, term, as_path: f'delete policy-options policy-statement {policy.name} term {term} from as-path {as_path}', PolicyFeatures.FROM_AS_PATH_GROUP: lambda policy, term, group: f'delete policy-options policy-statement {policy.name} term {term} from as-path-group {group}', PolicyFeatures.FROM_COLOR: lambda policy, term, color: f'delete policy-options policy-statement {policy.name} term {term} from color {color}', PolicyFeatures.FROM_COMMUNITY: lambda policy, term, community: f'delete policy-options policy-statement {policy.name} term {term} from community {community}', PolicyFeatures.FROM_FAMILY: lambda policy, term, family: f'delete policy-options policy-statement {policy.name} term {term} from family {family}', PolicyFeatures.FROM_INSTANCE: lambda: f'', PolicyFeatures.FROM_INTERFACE: lambda policy, term, interface: f'delete policy-options policy-statement {policy.name} term {term} from interface {interface.name}', PolicyFeatures.FROM_LEVEL: lambda: f'', PolicyFeatures.FROM_LOCAL_PREFERENCE: lambda policy, term, preference: f'delete policy-options policy-statement {policy.name} term {term} from local-preference {preference}', PolicyFeatures.FROM_METRIC: lambda policy, term, metric: f'delete policy-options policy-statement {policy.name} term {term} from metric {metric}', PolicyFeatures.FROM_NEIGHBOUR: lambda policy, term, neighbour: f'delete policy-options policy-statement {policy.name} term {term} from neighbor {int_to_ip(neighbour.address)}', PolicyFeatures.FROM_ORIGIN: lambda policy, term, origin: f'delete policy-options policy-statement {policy.name} term {term} from origin {origin}', PolicyFeatures.FROM_POLICY: lambda policy, term, policy2: f'delete policy-options policy-statement {policy.name} term {term} from policy {policy2}', PolicyFeatures.FROM_PREFIX_LIST: lambda policy, term, prefix_list: f'delete policy-options policy-statement {policy.name} term {term} from prefix-list {prefix_list}', PolicyFeatures.FROM_PREFIX_LIST_FILTER: lambda policy, term, prefix_list, match_type: f'delete policy-options policy-statement {policy.name} term {term} from prefix-list-filter {prefix_list} {match_type}', PolicyFeatures.FROM_PROTOCOL: lambda policy, term, protocol: f'delete policy-options policy-statement {policy.name} term {term} from protocol {protocol}', PolicyFeatures.FROM_RIB: lambda policy, term, rib: f'delete policy-options policy-statement {policy.name} term {term} from rib {rib}', PolicyFeatures.FROM_ROUTE_FILTER: lambda policy, term, net, match_type: f'delete policy-options policy-statement {policy.name} term {term} from route-filter {int_to_ip(net.address)}/{net.prefix} {match_type}', PolicyFeatures.FROM_ROUTE_TYPE: lambda policy, term, route_type: f'delete policy-options policy-statement {policy.name} term {term} from route-type {route_type}', PolicyFeatures.FROM_SOURCE_ADDRESS_FILTER: lambda policy, term, source_filter, match_type: f'delete policy-options policy-statement {policy.name} term {term} from source-address-filter {source_filter} {match_type}', PolicyFeatures.FROM_TAG: lambda policy, term, tag: f'delete policy-options policy-statement {policy.name} term {term} from tag {tag}', PolicyFeatures.FROM_NEXT_HOP: lambda policy, term, next_hop: f'delete policy-options policy-statement {policy.name} term {term} from next-hop {int_to_ip(next_hop)}', PolicyFeatures.TO_LEVEL: lambda: f'', PolicyFeatures.TO_RIB: lambda: f'', PolicyFeatures.THEN_ACCEPT: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then accept', PolicyFeatures.THEN_AS_PATH_EXPAND: lambda policy, term, n: f'delete policy-options policy-statement {policy.name} term {term} then as-path-expand last-as count {n}', PolicyFeatures.THEN_AS_PATH_PREPEND: lambda policy, term, as_path: f'delete policy-options policy-statement {policy.name} term {term} then as-path-prepend {as_path}', PolicyFeatures.THEN_COLOR: lambda policy, term, add, color: f'delete policy-options policy-statement {policy.name} term {term} then color {add} {color}', PolicyFeatures.THEN_COLOR2: lambda policy, term, add, color: f'delete policy-options policy-statement {policy.name} term {term} then color2 {add} {color}', PolicyFeatures.THEN_COMMUNITY_ADD: lambda policy, term, community: f'delete policy-options policy-statement {policy.name} term {term} then community add {community}', PolicyFeatures.THEN_COMMUNITY_DELETE: lambda policy, term, community: f'delete policy-options policy-statement {policy.name} term {term} then community delete {community}', PolicyFeatures.THEN_COMMUNITY_SET: lambda policy, term, community: f'delete policy-options policy-statement {policy.name} term {term} then community set {community}', PolicyFeatures.THEN_COS_NEXT_HOP_MAP: lambda policy, term, cos_map: f'delete policy-options policy-statement {policy.name} term {term} then cos-next-hop-map {cos_map}', PolicyFeatures.THEN_DEFAULT_ACTION_ACCEPT: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then default-action accept', PolicyFeatures.THEN_DEFAULT_ACTION_REJECT: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then default-action reject', PolicyFeatures.THEN_EXTERNAL: lambda policy, term, metric_type: f'delete policy-options policy-statement {policy.name} term {term} then external type {metric_type}', PolicyFeatures.THEN_FORWARDING_CLASS: lambda policy, term, fwd_class: f'delete policy-options policy-statement {policy.name} term {term} then forwarding-class {fwd_class}', PolicyFeatures.THEN_INSTALL_NEXTHOP: lambda: f'', PolicyFeatures.THEN_LOCAL_PREFERENCE: lambda policy, term, pref: f'delete policy-options policy-statement {policy.name} term {term} then local-preference {pref}', PolicyFeatures.THEN_METRIC: lambda policy, term, metric: f'delete policy-options policy-statement {policy.name} term {term} then metric {metric}', PolicyFeatures.THEN_METRIC_ADD: lambda policy, term, metric: f'delete policy-options policy-statement {policy.name} term {term} then metric add {metric}', PolicyFeatures.THEN_METRIC_EXPRESSION: lambda: f'', PolicyFeatures.THEN_METRIC_IGP: lambda policy, term, offset: f'delete policy-options policy-statement {policy.name} term {term} then metric igp {offset}', PolicyFeatures.THEN_METRIC2: lambda policy, term, metric: f'delete policy-options policy-statement {policy.name} term {term} then metric2 {metric}', PolicyFeatures.THEN_METRIC2_EXPRESSION: lambda: f'', PolicyFeatures.THEN_NEXT_HOP: lambda policy, term, next_hop: f'delete policy-options policy-statement {policy.name} term {term} then next-hop {int_to_ip(next_hop)}', PolicyFeatures.THEN_NEXT_HOP_SELF: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then next-hop self', PolicyFeatures.THEN_NEXT_POLICY: lambda: f'', PolicyFeatures.THEN_NEXT_TERM: lambda: f'', PolicyFeatures.THEN_ORIGIN: lambda policy, term, origin: f'delete policy-options policy-statement {policy.name} term {term} then origin {origin}', PolicyFeatures.THEN_PREFERENCE: lambda policy, term, pref: f'delete policy-options policy-statement {policy.name} term {term} then preference {pref}', PolicyFeatures.THEN_PRIORITY: lambda policy, term, priority: f'delete policy-options policy-statement {policy.name} term {term} then priority {priority}', PolicyFeatures.THEN_REJECT: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then reject', PolicyFeatures.THEN_TAG: lambda policy, term, tag: f'delete policy-options policy-statement {policy.name} term {term} then tag {tag}', PolicyFeatures.POLICY_MATCH_FEATURE_BGP_OUT: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args), PolicyFeatures.POLICY_MATCH_FEATURE_BGP_IN: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args), PolicyFeatures.POLICY_MATCH_FEATURE_OSPF_OUT: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args), PolicyFeatures.POLICY_MATCH_FEATURE_OSPF_IN: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args), PolicyFeatures.POLICY_SET_FEATURE_BGP_OUT: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args), PolicyFeatures.POLICY_SET_FEATURE_BGP_IN: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args), PolicyFeatures.POLICY_SET_FEATURE_OSPF_OUT: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args), PolicyFeatures.POLICY_SET_FEATURE_OSPF_IN: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args) } def config_mode(router, feature, arg): return () def exit_config_mode(feature): return [] def generate_maps_lists_config(router): config = [] for as_path in router.as_path_lists: config.append(f'set policy-options as-path {as_path.name} {as_path.regex}') for community in router.comm_lists: config.append(f'set policy-options community {community.name} members [{",".join(community.comms)}]') for prefix_list in router.prefix_lists: for seq in prefix_list.prefix: net = prefix_list.prefix[seq] config.append(f'set policy-options prefix-list {prefix_list.name} {int_to_ip(net.address)}/{net.prefix}') for policy in router.ospf_in_route_maps: config.append(f'set policy-options policy-statement {policy.name} then accept') for policy in router.ospf_out_route_maps: config.append(f'set policy-options policy-statement {policy.name} then accept') for policy in router.bgp_in_route_maps: config.append(f'set policy-options policy-statement {policy.name} then accept') for policy in router.bgp_out_route_maps: config.append(f'set policy-options policy-statement {policy.name} then accept') return config def generate_bgp_config(router): config = [ f'set routing-options autonomous-system {router.AS.num}', f'set routing-options router-id {int_to_ip(router.router_id)}', f'set protocols bgp group IBGP type internal', f'set protocols bgp group IBGP local-address {int_to_ip(router.router_id)}', f'set protocols bgp group EBGP type external' ] for neighbour in router.bgp_neighbours: if neighbour.AS == router.AS: config.append(f'set protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)}') else: config.append( f'set protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} peer-as {neighbour.AS.num}') return config def get_base_config(router): config = [ f'set system host-name {router.name}', f'set interfaces lo0.0 family inet address {int_to_ip(router.router_id)}/32' ] for interface in router.interfaces: if interface.address is not None: config.append(f'set interfaces {interface.name} family inet address ' f'{int_to_ip(interface.address)}/{interface.prefix}') if interface.area is not None: config.append(f'set protocols ospf area {interface.area} interface {interface.name}') for network, interface in router.fixed_static_routes: config.append(f'set routing-options static route {int_to_ip(network.address)}/{network.prefix} next-hop {interface.name}') for area in router.ospf_areas: if area.type == OSPF_Area_Type.STUB: config.append(f'set protocols ospf area {area} stub') elif area.type == OSPF_Area_Type.NSSA: config.append(f'set protocols ospf area {area} nssa') for protocol in ['direct', 'static', 'ospf', 'bgp']: config.append(f'set policy-options policy-statement send-{protocol} term 1 from protocol {protocol}') config.append(f'set policy-options policy-statement send-{protocol} term 1 then accept') if Protocols.BGP in router.enabled_protocols: config.append(f'set protocols bgp hold-time 3') config.extend(generate_bgp_config(router)) config.extend(generate_maps_lists_config(router)) return config def write_config(router, path): config = get_base_config(router) os.makedirs(os.path.dirname(path), exist_ok=True) with open(f'{path}{router.name}.set', 'w') as f: f.write('\n'.join(config))
89.486413
219
0.745832
4,332
32,931
5.543167
0.045937
0.051639
0.071961
0.114271
0.94453
0.925457
0.91388
0.890684
0.886853
0.852622
0
0.001431
0.151286
32,931
367
220
89.730245
0.857741
0
0
0.169753
0
0.317901
0.455498
0.028393
0
0
0
0
0
1
0.018519
false
0.018519
0.030864
0.006173
0.064815
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
2a34900f1755015a41fdc6669c2a89e96570d4f2
9,790
py
Python
tests/test_selective_extraction.py
alizab223/swish-utilities
cc3c30c3aad4fd40f7a72f939b92234d3acf4be2
[ "Apache-2.0" ]
null
null
null
tests/test_selective_extraction.py
alizab223/swish-utilities
cc3c30c3aad4fd40f7a72f939b92234d3acf4be2
[ "Apache-2.0" ]
null
null
null
tests/test_selective_extraction.py
alizab223/swish-utilities
cc3c30c3aad4fd40f7a72f939b92234d3acf4be2
[ "Apache-2.0" ]
null
null
null
import json from unittest import TestCase from click.testing import CliRunner from pandas.io.parsers import read_csv import requests_mock from run import cli, main from tests.common import SNOW_RESPONSE1, SNOW_RESPONSE_WITH_CUSTOM_ID, UNITEST_OUTPUT_FILE, UNITEST_OUTPUT_FILE_CSV, UNITEST_OUTPUT_FILE_PREFIX, patch_for_tests import os import os.path import csv patch_for_tests() TEST_IDS_FILE = "tests/data/sys_ids_list.csv" TEST_IDS_FILE2 = "tests/data/custom_ids_list.csv" class SelectiveExtractionTesting(TestCase): def setUp(self) -> None: if os.path.isfile(UNITEST_OUTPUT_FILE): os.remove(UNITEST_OUTPUT_FILE) if os.path.isfile(UNITEST_OUTPUT_FILE_CSV): os.remove(UNITEST_OUTPUT_FILE_CSV) def test_no_filter(self): assert not os.path.isfile( UNITEST_OUTPUT_FILE), "The file should be deleted" mock_session = requests_mock.Mocker() mock_session.register_uri(requests_mock.ANY, 'https://dev71074.service-now.com/api/now/table/sys_audit', text=SNOW_RESPONSE1) mock_session.start() args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident", "--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000", "--start_date", "2021-10-03", "--end_date", "2021-10-04"] runner = CliRunner() result = runner.invoke(cli, args, catch_exceptions=False) print(result.output) assert result.exit_code == 0 assert os.path.isfile(UNITEST_OUTPUT_FILE), "No output file found" data = [] with open(UNITEST_OUTPUT_FILE, 'r') as f: data = json.load(f) assert len([d for d in data if d['sys_id'] == '669009b4874330105fd965f73cbb3533']) assert len([d for d in data if d['sys_id'] == '123']) assert len([d for d in data if d['sys_id'] == '456']) assert not len([d for d in data if d['sys_id'] == '567']) def test_no_filter_to_csv(self): assert not os.path.isfile( UNITEST_OUTPUT_FILE_CSV), "The file should be deleted" mock_session = requests_mock.Mocker() mock_session.register_uri(requests_mock.ANY, 'https://dev71074.service-now.com/api/now/table/sys_audit', text=SNOW_RESPONSE1) mock_session.start() args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident", "--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000", "--start_date", "2021-10-03", "--end_date", "2021-10-04", "--output_format", "csv"] runner = CliRunner() result = runner.invoke(cli, args, catch_exceptions=False) print(result.output) assert result.exit_code == 0 assert os.path.isfile(UNITEST_OUTPUT_FILE_CSV), "No output file found" data = [] with open(UNITEST_OUTPUT_FILE_CSV, 'r') as f: data = list(csv.DictReader(f)) assert len([d for d in data if d['sys_id'] == '669009b4874330105fd965f73cbb3533']) assert len([d for d in data if d['sys_id'] == '123']) assert len([d for d in data if d['sys_id'] == '456']) assert not len([d for d in data if d['sys_id'] == '567']) def test_selective(self): assert not os.path.isfile( UNITEST_OUTPUT_FILE), "The file should be deleted" mock_session = requests_mock.Mocker() mock_session.register_uri(requests_mock.ANY, 'https://dev71074.service-now.com/api/now/table/sys_audit', text=SNOW_RESPONSE1) mock_session.start() args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident", "--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000", "--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", TEST_IDS_FILE] runner = CliRunner() result = runner.invoke(cli, args, catch_exceptions=False) print(result.output) assert result.exit_code == 0 assert os.path.isfile( UNITEST_OUTPUT_FILE), "No output file found" data = [] with open(UNITEST_OUTPUT_FILE, 'r') as f: data = json.load(f) assert len([d for d in data if d['sys_id'] == '669009b4874330105fd965f73cbb3533']) == 0 ids = read_csv(TEST_IDS_FILE, encoding='utf-8')['sys_id'] for id in ids: assert len([d for d in data if d['sys_id'] == str(id)]) def test_selective_with_prop_extraction(self): assert not os.path.isfile( UNITEST_OUTPUT_FILE), "The file should be deleted" mock_session = requests_mock.Mocker() mock_session.register_uri(requests_mock.ANY, 'https://dev71074.service-now.com/api/now/table/sys_audit', text=SNOW_RESPONSE1) mock_session.start() args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident", "--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000", "--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", TEST_IDS_FILE, "--out_props_csv_path", "qwerty.csv", "--out_prop_name", "sys_created_on"] runner = CliRunner() result = runner.invoke(cli, args, catch_exceptions=False) print(result.output) assert result.exit_code == 0 data = read_csv(f"qwerty.csv", encoding='utf-8') assert len(data['sys_created_on'].values) == 1, "wrong output count" def test_selective_with_id_column(self): assert not os.path.isfile( UNITEST_OUTPUT_FILE), "The file should be deleted" mock_session = requests_mock.Mocker() mock_session.register_uri(requests_mock.ANY, 'https://dev71074.service-now.com/api/now/table/sys_audit', text=SNOW_RESPONSE_WITH_CUSTOM_ID) mock_session.start() args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident", "--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000", "--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", TEST_IDS_FILE2, "--id_field_name", "custom_sys_id"] runner = CliRunner() result = runner.invoke(cli, args, catch_exceptions=False) print(result.output) assert result.exit_code == 0 assert os.path.isfile(UNITEST_OUTPUT_FILE), "No output file found" data = [] with open(UNITEST_OUTPUT_FILE, 'r') as f: data = json.load(f) assert len([d for d in data if d['custom_sys_id'] == '669009b4874330105fd965f73cbb3533']) == 0 ids = read_csv(TEST_IDS_FILE2, encoding='utf-8')['custom_sys_id'] for id in ids: assert len([d for d in data if d['custom_sys_id'] == str(id)]) def test_selective_with_data_id_name(self): assert not os.path.isfile( UNITEST_OUTPUT_FILE), "The file should be deleted" mock_session = requests_mock.Mocker() mock_session.register_uri(requests_mock.ANY, 'https://dev71074.service-now.com/api/now/table/sys_audit', text=SNOW_RESPONSE_WITH_CUSTOM_ID) mock_session.start() args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident", "--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000", "--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", TEST_IDS_FILE2, "--id_field_name", "custom_sys_id", "--data_id_name", "custom_sys_id"] runner = CliRunner() result = runner.invoke(cli, args, catch_exceptions=False) print(result.output) assert result.exit_code == 0 assert os.path.isfile(UNITEST_OUTPUT_FILE), "No output file found" data = [] with open(UNITEST_OUTPUT_FILE, 'r') as f: data = json.load(f) assert len([d for d in data if d['custom_sys_id'] == '669009b4874330105fd965f73cbb3533']) == 0 ids = read_csv(TEST_IDS_FILE, encoding='utf-8')['sys_id'] for id in ids: assert len([d for d in data if d['custom_sys_id'] == str(id)]) def test_file_doesnt_exist(self): args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident", "--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000", "--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", "qwqweqweqweqew.txt"] runner = CliRunner() result = runner.invoke(cli, args, catch_exceptions=False) print(result.output) assert result.exit_code == 0 assert not os.path.isfile( UNITEST_OUTPUT_FILE), "The file should not exist"
47.067308
160
0.60143
1,252
9,790
4.46885
0.115016
0.051832
0.072922
0.047542
0.866488
0.853262
0.853262
0.840572
0.83378
0.826273
0
0.057064
0.258938
9,790
207
161
47.294686
0.714128
0
0
0.735294
0
0.041176
0.288355
0.022165
0
0
0
0
0.2
1
0.047059
false
0.041176
0.058824
0
0.111765
0.041176
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
2a994de539e2b5e6ddbcd4033437ddb17e4fe853
5,297
py
Python
main.py
Kubik-Kubik/cs-go-map-chooser
82e2e27999d151a06777fc3ed9439a52a2206568
[ "Unlicense" ]
null
null
null
main.py
Kubik-Kubik/cs-go-map-chooser
82e2e27999d151a06777fc3ed9439a52a2206568
[ "Unlicense" ]
null
null
null
main.py
Kubik-Kubik/cs-go-map-chooser
82e2e27999d151a06777fc3ed9439a52a2206568
[ "Unlicense" ]
null
null
null
from PyQt5 import QtWidgets, uic import sys from ui import Ui_Dialog from PyQt5 import QtCore, QtGui, QtWidgets import sys from ui import Ui_Dialog # create app app = QtWidgets.QApplication(sys.argv) # create dialog Dialog = QtWidgets.QDialog() ui = Ui_Dialog() ui.setupUi(Dialog) Dialog.setWindowTitle("Map Picker") Dialog.setWindowIcon(QtGui.QIcon("logo.png")) Dialog.show() # hook logic def reset(): ui.vertigo_img.setStyleSheet("QPushButton {background-image: url(images/vertigo.png);" "border: none;}") ui.inferno_img.setStyleSheet("QPushButton {background-image: url(images/inferno.png);" "border: none;}") ui.overpass_img.setStyleSheet("QPushButton {background-image: url(images/overpass.png);" "border: none;}") ui.cobble_img.setStyleSheet("QPushButton {background-image: url(images/cobble.png);" "border: none;}") ui.train_img.setStyleSheet("QPushButton {background-image: url(images/train.png);" "border: none;}") ui.dust_img.setStyleSheet("QPushButton {background-image: url(images/dust.png);" "border: none;}") ui.nuke_img.setStyleSheet("QPushButton {background-image: url(images/nuke.png);" "border: none;}") ui.realty_img.setStyleSheet("QPushButton {background-image: url(images/rialto.png);" "border: none;}") ui.lake_img.setStyleSheet("QPushButton {background-image: url(images/lake.png);" "border: none;}") def ban_map(): if ui.vertigo_btn.isChecked(): ui.vertigo_img.setStyleSheet("QPushButton {background-image: url(images/vertigo_ban.png);" "border: none;}") elif ui.inferno_btn.isChecked(): ui.inferno_img.setStyleSheet("QPushButton {background-image: url(images/inferno_ban.png);" "border: none;}") elif ui.overpass_btn.isChecked(): ui.overpass_img.setStyleSheet("QPushButton {background-image: url(images/overpass_ban.png);" "border: none;}") elif ui.cobble_btn.isChecked(): ui.cobble_img.setStyleSheet("QPushButton {background-image: url(images/cobble_ban.png);" "border: none;}") elif ui.train_btn.isChecked(): ui.train_img.setStyleSheet("QPushButton {background-image: url(images/train_ban.png);" "border: none;}") elif ui.dust_btn.isChecked(): ui.dust_img.setStyleSheet("QPushButton {background-image: url(images/dust_ban.png);" "border: none;}") elif ui.nuke_btn.isChecked(): ui.nuke_img.setStyleSheet("QPushButton {background-image: url(images/nuke_ban.png);" "border: none;}") elif ui.rialto_btn.isChecked(): ui.realty_img.setStyleSheet("QPushButton {background-image: url(images/rialto_ban.png);" "border: none;}") elif ui.lake_btn.isChecked(): ui.lake_img.setStyleSheet("QPushButton {background-image: url(images/lake_ban.png);" "border: none;}") def pick_map(): if ui.vertigo_btn.isChecked(): ui.vertigo_img.setStyleSheet("QPushButton {background-image: url(images/vertigo_pick.png);" "border: none;}") elif ui.inferno_btn.isChecked(): ui.inferno_img.setStyleSheet("QPushButton {background-image: url(images/inferno_pick.png);" "border: none;}") elif ui.overpass_btn.isChecked(): ui.overpass_img.setStyleSheet("QPushButton {background-image: url(images/overpass_pick.png);" "border: none;}") elif ui.cobble_btn.isChecked(): ui.cobble_img.setStyleSheet("QPushButton {background-image: url(images/cobble_pick.png);" "border: none;}") elif ui.train_btn.isChecked(): ui.train_img.setStyleSheet("QPushButton {background-image: url(images/train_pick.png);" "border: none;}") elif ui.dust_btn.isChecked(): ui.dust_img.setStyleSheet("QPushButton {background-image: url(images/dust_pick.png);" "border: none;}") elif ui.nuke_btn.isChecked(): ui.nuke_img.setStyleSheet("QPushButton {background-image: url(images/nuke_pick.png);" "border: none;}") elif ui.rialto_btn.isChecked(): ui.realty_img.setStyleSheet("QPushButton {background-image: url(images/rialto_pick.png);" "border: none;}") elif ui.lake_btn.isChecked(): ui.lake_img.setStyleSheet("QPushButton {background-image: url(images/lake_pick.png);" "border: none;}") ui.ban_btn.clicked.connect(ban_map) ui.pick_btn.clicked.connect(pick_map) ui.reset_btn.clicked.connect(reset) # run main loop sys.exit(app.exec_())
47.294643
102
0.58316
551
5,297
5.471869
0.108893
0.143284
0.241791
0.331343
0.824876
0.824876
0.806302
0.787065
0.787065
0.787065
0
0.000532
0.289975
5,297
111
103
47.720721
0.801117
0.009251
0
0.532609
0
0
0.375292
0.135035
0
0
0
0
0
1
0.032609
false
0.054348
0.065217
0
0.097826
0
0
0
0
null
0
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
1
0
0
0
0
0
9
aa787d3354b46adc19835d621fb2b7e7853329aa
3,701
py
Python
events/migrations/0067_modify_several_fields_on_delete.py
jannetasa/linkedevents
0d72c3266025b1bd71b3f328fd2f8a4b75e1b82b
[ "MIT" ]
20
2015-05-28T16:02:00.000Z
2021-07-14T06:36:19.000Z
events/migrations/0067_modify_several_fields_on_delete.py
jannetasa/linkedevents
0d72c3266025b1bd71b3f328fd2f8a4b75e1b82b
[ "MIT" ]
358
2015-02-04T09:07:19.000Z
2022-03-28T12:10:22.000Z
events/migrations/0067_modify_several_fields_on_delete.py
jannetasa/linkedevents
0d72c3266025b1bd71b3f328fd2f8a4b75e1b82b
[ "MIT" ]
38
2015-02-23T13:16:02.000Z
2021-12-13T07:48:23.000Z
# Generated by Django 2.2.9 on 2020-01-08 08:22 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('events', '0066_modify_datasource_owner_on_delete'), ] operations = [ migrations.AlterField( model_name='event', name='created_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_event_created_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='event', name='last_modified_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_event_modified_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='image', name='created_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='EventImage_created_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='image', name='last_modified_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='EventImage_last_modified_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='image', name='license', field=models.ForeignKey(default='cc_by', null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='images', to='events.License', verbose_name='License'), ), migrations.AlterField( model_name='keyword', name='created_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_keyword_created_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='keyword', name='last_modified_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_keyword_modified_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='keywordset', name='created_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_keywordset_created_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='keywordset', name='last_modified_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_keywordset_modified_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='place', name='created_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_place_created_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='place', name='last_modified_by', field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_place_modified_by', to=settings.AUTH_USER_MODEL), ), migrations.AlterField( model_name='place', name='replaced_by', field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='aliases', to='events.Place'), ), ]
48.697368
180
0.668738
444
3,701
5.304054
0.128378
0.047558
0.077282
0.121444
0.850955
0.850955
0.816136
0.816136
0.816136
0.816136
0
0.006531
0.213996
3,701
75
181
49.346667
0.803025
0.012159
0
0.666667
1
0
0.155172
0.078818
0
0
0
0
0
1
0
false
0
0.043478
0
0.086957
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
aaa2cbabba118d0912358463f786dc71ba0680b1
133
py
Python
hypothalamus_seg/__init__.py
BBillot/hypothlamus_seg
a97bf109370ce758ad2ee457f12ba173ba7da818
[ "Apache-2.0" ]
21
2020-05-27T10:11:45.000Z
2022-03-12T20:19:22.000Z
hypothalamus_seg/__init__.py
BBillot/hypothlamus_seg
a97bf109370ce758ad2ee457f12ba173ba7da818
[ "Apache-2.0" ]
5
2020-01-28T22:12:30.000Z
2022-03-03T15:13:37.000Z
hypothalamus_seg/__init__.py
BBillot/hypothlamus_seg
a97bf109370ce758ad2ee457f12ba173ba7da818
[ "Apache-2.0" ]
5
2021-05-13T06:31:36.000Z
2022-03-13T19:11:58.000Z
from . import augmentation_model from . import model_inputs from . import metrics_model from . import predict from . import training
22.166667
32
0.81203
18
133
5.833333
0.444444
0.47619
0.285714
0
0
0
0
0
0
0
0
0
0.150376
133
5
33
26.6
0.929204
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
2aee171d72ccbeed70c5a6947afde318c51db467
40
py
Python
OMASS4/Opt_SNR_AR/__init__.py
DBernardes/OMASS4
30d2edc961463253cc120bc8ca1d74a0a73d922d
[ "MIT" ]
null
null
null
OMASS4/Opt_SNR_AR/__init__.py
DBernardes/OMASS4
30d2edc961463253cc120bc8ca1d74a0a73d922d
[ "MIT" ]
null
null
null
OMASS4/Opt_SNR_AR/__init__.py
DBernardes/OMASS4
30d2edc961463253cc120bc8ca1d74a0a73d922d
[ "MIT" ]
null
null
null
from .optimize_snr_ar import Opt_SNR_AR
20
39
0.875
8
40
3.875
0.75
0.322581
0
0
0
0
0
0
0
0
0
0
0.1
40
1
40
40
0.861111
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
1
0
null
1
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
63077cb6d7c7f4ae64f4badcabcc599b41c0d809
57,404
py
Python
mseg/label_preparation/dataset_update_records.py
khose/mseg-api
ae649c40918a5fc8a14a88a12b01b8032cc53492
[ "CC-BY-4.0" ]
213
2020-04-25T02:51:42.000Z
2022-03-25T09:57:26.000Z
mseg/label_preparation/dataset_update_records.py
khose/mseg-api
ae649c40918a5fc8a14a88a12b01b8032cc53492
[ "CC-BY-4.0" ]
19
2020-04-28T05:47:50.000Z
2022-03-18T03:05:07.000Z
mseg/label_preparation/dataset_update_records.py
khose/mseg-api
ae649c40918a5fc8a14a88a12b01b8032cc53492
[ "CC-BY-4.0" ]
35
2020-04-29T16:11:43.000Z
2022-03-28T15:50:15.000Z
#!/usr/bin/python3 from mseg.label_preparation.relabeled_data_containers import DatasetClassUpdateRecord """ COCO Panoptic""" cocop_update_records = [ # cabinet-merged-classification,fence-guardrail-classification # person_rider_classification,rug-merged,showercurtain_curtainother_classification # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'desk', 'cocop_cabinet-merged/cocopanoptic_desk_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'bookshelf', 'cocop_cabinet-merged/cocopanoptic_bookshelf_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'bookshelf', 'cocop_cabinet-merged/cocopanoptic_bookshelf_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'cabinet', 'cocop_cabinet-merged/cocopanoptic_cabinet_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'cabinet', 'cocop_cabinet-merged/cocopanoptic_cabinet_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'unlabeled', 'cocop_cabinet-merged/cocopanoptic_cabinetmerged_unlabel_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'unlabeled', 'cocop_cabinet-merged/cocopanoptic_cabinetmerged_unlabel_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'counter_other', 'cocop_cabinet-merged/cocopanoptic_counter_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'chest_of_drawers', 'cocop_cabinet-merged/cocopanoptic_dresser_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'chest_of_drawers', 'cocop_cabinet-merged/cocopanoptic_dresser_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'nightstand', 'cocop_cabinet-merged/cocopanoptic_nightstand_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'nightstand', 'cocop_cabinet-merged/cocopanoptic_nightstand_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'wardrobe', 'cocop_cabinet-merged/cocopanoptic_wardrobe_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'fence', 'fence', 'cocop_fence-guardrail/cocopanoptic_fence_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'fence', 'fence', 'cocop_fence-guardrail/cocopanoptic_fence_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'fence', 'guard_rail', 'cocop_fence-guardrail/cocopanoptic_guardrail_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'fence', 'guard_rail', 'cocop_fence-guardrail/cocopanoptic_guardrail_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'rug_floormat', 'unlabeled', 'cocop_rug-merged/cocopanoptic_rug-merged_unlabel-list_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'rug_floormat', 'unlabeled', 'cocop_rug-merged/cocopanoptic_rug-merged_unlabel-list_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'rug_floormat', 'floor', 'cocop_rug-merged/floor-carpet_80percent_confidence_cocopanoptic_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'rug_floormat', 'floor', 'cocop_rug-merged/floor-carpet_80percent_confidence_cocopanoptic_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'rug_floormat', 'rug_floormat', 'cocop_rug-merged/rug-floormat_80percent_confidence_cocopanoptic_train.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'rug_floormat', 'rug_floormat', 'cocop_rug-merged/rug-floormat_80percent_confidence_cocopanoptic_val.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'curtain_other', 'curtain_other', 'cocop_curtain/2019_08_16_coco_other-curtain_80percent_conf_valsplit.txt'), DatasetClassUpdateRecord('cocopanoptic', 'val', 'curtain_other', 'shower_curtain', 'cocop_curtain/2019_08_16_coco_showercurtain_80percent_conf_valsplit.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'curtain_other', 'curtain_other', 'cocop_curtain/2019_08_17_coco_other-curtain_train_80percent_conf.txt'), DatasetClassUpdateRecord('cocopanoptic', 'train', 'curtain_other', 'shower_curtain', 'cocop_curtain/2019_08_17_coco_showercurtain_train_80percent_conf.txt'), # cocop_cocop_chair (COCOP CHAIR-> UNIVERSAL CHAIR-OTHER ) # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'armchair', 'cocop_chair/cocop_train_cocop_chair_to_armchair.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'basket', 'cocop_chair/cocop_train_cocop_chair_to_basket.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'bench', 'cocop_chair/cocop_train_cocop_chair_to_bench.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'chair_other', 'cocop_chair/cocop_train_cocop_chair_to_chair-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'unlabeled', 'cocop_chair/cocop_train_cocop_chair_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'ottoman', 'cocop_chair/cocop_train_cocop_chair_to_ottoman.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'seat', 'cocop_chair/cocop_train_cocop_chair_to_seat.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'slow_wheeled_object', 'cocop_chair/cocop_train_cocop_chair_to_slowwheeledobject.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'stool', 'cocop_chair/cocop_train_cocop_chair_to_stool.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'swivel_chair', 'cocop_chair/cocop_train_cocop_chair_to_swivel-chair.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'armchair', 'cocop_chair/cocop_val_cocop_chair_to_armchair.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'chair_other', 'cocop_chair/cocop_val_cocop_chair_to_chair-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'unlabeled', 'cocop_chair/cocop_val_cocop_chair_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'ottoman', 'cocop_chair/cocop_val_cocop_chair_to_ottoman.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'seat', 'cocop_chair/cocop_val_cocop_chair_to_seat.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'slow_wheeled_object', 'cocop_chair/cocop_val_cocop_chair_to_slowwheeledobject.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'stool', 'cocop_chair/cocop_val_cocop_chair_to_stool.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'swivel_chair', 'cocop_chair/cocop_val_cocop_chair_to_swivel-chair.txt'), # cocop_cocop_counter (COCO COUNTER -> UNIVERSAL COUNTER-OTHER) DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'bathroom_counter', 'cocop_cocop_counter/cocop_train_cocop_counter_to_bathroom-counter.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'counter_other', 'cocop_cocop_counter/cocop_train_cocop_counter_to_counter-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'kitchen_island', 'cocop_cocop_counter/cocop_train_cocop_counter_to_kitchen-island.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'nightstand', 'cocop_cocop_counter/cocop_train_cocop_counter_to_nightstand.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'unlabeled', 'cocop_cocop_counter/cocop_train_cocop_counter_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'counter_other', 'bathroom_counter', 'cocop_cocop_counter/cocop_val_cocop_counter_to_bathroom-counter.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'counter_other', 'counter_other', 'cocop_cocop_counter/cocop_val_cocop_counter_to_counter-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'counter_other', 'kitchen_island', 'cocop_cocop_counter/cocop_val_cocop_counter_to_kitchen-island.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'counter_other', 'unlabeled', 'cocop_cocop_counter/cocop_val_cocop_counter_to_None_of_these.txt'), # cocop_cocop_diningtable (COCOP {table-merged, dining table} -> UNIVERSAL TABLE ) DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'bathroom_counter', 'cocop_diningtable/cocop_train_cocop_diningtable_to_bathroom-counter.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'box', 'cocop_diningtable/cocop_train_cocop_diningtable_to_box.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'chest_of_drawers', 'cocop_diningtable/cocop_train_cocop_diningtable_to_chestofdrawers.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'counter_other', 'cocop_diningtable/cocop_train_cocop_diningtable_to_counter-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'desk', 'cocop_diningtable/cocop_train_cocop_diningtable_to_desk-and-table.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'desk', 'cocop_diningtable/cocop_train_cocop_diningtable_to_desk.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'kitchen_island', 'cocop_diningtable/cocop_train_cocop_diningtable_to_kitchen-island.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'nightstand', 'cocop_diningtable/cocop_train_cocop_diningtable_to_nightstand.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'unlabeled', 'cocop_diningtable/cocop_train_cocop_diningtable_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'table', 'cocop_diningtable/cocop_train_cocop_diningtable_to_table.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'box', 'cocop_diningtable/cocop_val_diningtable_to_box.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'counter_other', 'cocop_diningtable/cocop_val_diningtable_to_counterother.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'desk', 'cocop_diningtable/cocop_val_diningtable_to_desk.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'desk', 'cocop_diningtable/cocop_val_diningtable_to_deskandtable.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'kitchen_island', 'cocop_diningtable/cocop_val_diningtable_to_kitchenisland.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'table', 'cocop_diningtable/cocop_val_diningtable_to_table.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'unlabeled', 'cocop_diningtable/cocop_val_diningtable_to_unlabel.txt'), # cocop_cocop_table-merged DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'base', 'cocop_table-merged/cocop_train_cocop_table-merged_to_base.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'bathroom_counter','cocop_table-merged/cocop_train_cocop_table-merged_to_bathroom-counter.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'cabinet', 'cocop_table-merged/cocop_train_cocop_table-merged_to_cabinet.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'chest_of_drawers','cocop_table-merged/cocop_train_cocop_table-merged_to_chestofdrawers.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'counter_other', 'cocop_table-merged/cocop_train_cocop_table-merged_to_counter-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'desk', 'cocop_table-merged/cocop_train_cocop_table-merged_to_desk.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'kitchen_island', 'cocop_table-merged/cocop_train_cocop_table-merged_to_kitchen-island.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'laptop', 'cocop_table-merged/cocop_train_cocop_table-merged_to_laptop.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'mountain_hill', 'cocop_table-merged/cocop_train_cocop_table-merged_to_mountainhill.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'nightstand', 'cocop_table-merged/cocop_train_cocop_table-merged_to_nightstand.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'unlabeled', 'cocop_table-merged/cocop_train_cocop_table-merged_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'pool_table', 'cocop_table-merged/cocop_train_cocop_table-merged_to_pooltable.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'stool', 'cocop_table-merged/cocop_train_cocop_table-merged_to_stool.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'desk', 'cocop_table-merged/cocop_train_cocop_table-merged_to_table-used-as-desk.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'table', 'cocop_table-merged/cocop_train_cocop_table-merged_to_table.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'bathroom_counter','cocop_table-merged/cocop_val_cocop_table-merged_to_bathroomcounter.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'counter_other', 'cocop_table-merged/cocop_val_cocop_table-merged_to_counterother.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'desk', 'cocop_table-merged/cocop_val_cocop_table-merged_to_desk.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'kitchen_island', 'cocop_table-merged/cocop_val_cocop_table-merged_to_kitchenisland.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'nightstand', 'cocop_table-merged/cocop_val_cocop_table-merged_to_nightstand.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'desk', 'cocop_table-merged/cocop_val_cocop_table-merged_to_table-used-as-desk.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'table', 'cocop_table-merged/cocop_val_cocop_table-merged_to_table.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'unlabeled', 'cocop_table-merged/cocop_val_cocop_table-merged_to_unlabel.txt'), # cocop_cocop_light (COCO LIGHT-> UNIVERSAL LIGHT-OTHER) DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'chandelier', 'cocop_light/cocop_train_cocop_light_to_chandelier.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'lamp', 'cocop_light/cocop_train_cocop_light_to_lamp.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'light_other', 'cocop_light/cocop_train_cocop_light_to_light-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'unlabeled', 'cocop_light/cocop_train_cocop_light_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'sconce', 'cocop_light/cocop_train_cocop_light_to_sconce.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'streetlight', 'cocop_light/cocop_train_cocop_light_to_streetlight.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'chandelier', 'cocop_light/cocop_val_cocop_light_to_chandelier.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'lamp', 'cocop_light/cocop_val_cocop_light_to_lamp.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'light_other', 'cocop_light/cocop_val_cocop_light_to_light-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'unlabeled', 'cocop_light/cocop_val_cocop_light_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'sconce', 'cocop_light/cocop_val_cocop_light_to_sconce.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'streetlight', 'cocop_light/cocop_val_cocop_light_to_streetlight.txt'), # cocopanoptic_cocop_motorcyclist_bicyclist (COCO PERSON -> UNIVERSAL PERSON-NON-RIDER) DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'bicyclist', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_bicyclist.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'motorcyclist', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_motorcyclist.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'unlabeled', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'person', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_person-non-rider.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'rider_other', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_rider-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'bicyclist', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_val_cocop_motorcyclist_bicyclist_to_bicyclist.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'motorcyclist', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_val_cocop_motorcyclist_bicyclist_to_motorcyclist.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'person', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_val_cocop_motorcyclist_bicyclist_to_person-non-rider.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'rider_other', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_val_cocop_motorcyclist_bicyclist_to_rider-other.txt'), # cocopanoptic_cocop_nonrider DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'bicyclist', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_bicyclist.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'motorcyclist', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_motorcyclist.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'unlabeled', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'person', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_person-non-rider.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'rider_other', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_rider-other.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'bicyclist', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_bicyclist.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'motorcyclist', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_motorcyclist.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'unlabeled', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_None_of_these.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'person', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_person-non-rider.txt'), DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'rider_other', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_rider-other.txt'), # Phase 3, cocop_waterother # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'fountain', 'cocop_waterother/cocop_train_cocop_waterother_to_fountain.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'unlabeled', 'cocop_waterother/cocop_train_cocop_waterother_to_None_of_these.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'playingfield', 'cocop_waterother/cocop_train_cocop_waterother_to_playingfield.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'river_lake', 'cocop_waterother/cocop_train_cocop_waterother_to_river-lake.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'sea', 'cocop_waterother/cocop_train_cocop_waterother_to_sea.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'swimming_pool', 'cocop_waterother/cocop_train_cocop_waterother_to_swimming-pool.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'terrain', 'cocop_waterother/cocop_train_cocop_waterother_to_terrain.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'wall', 'cocop_waterother/cocop_train_cocop_waterother_to_wall.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'water_other', 'cocop_waterother/cocop_train_cocop_waterother_to_water-other.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'waterfall', 'cocop_waterother/cocop_train_cocop_waterother_to_waterfall.txt'), DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'window', 'cocop_waterother/cocop_train_cocop_waterother_to_window.txt'), # Phase 3,cocop_runway_pavement # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocop','train', 'sidewalk_pavement', 'runway', 'cocop_runway_pavement/cocop_train_pavementmerged_to_airport_runway.txt'), DatasetClassUpdateRecord('cocop','train', 'sidewalk_pavement', 'floor', 'cocop_runway_pavement/cocop_train_pavementmerged_to_floor.txt'), DatasetClassUpdateRecord('cocop','train', 'sidewalk_pavement', 'sidewalk_pavement', 'cocop_runway_pavement/cocop_train_pavementmerged_to_pavementmerged.txt'), DatasetClassUpdateRecord('cocop','train', 'sidewalk_pavement', 'pier_wharf', 'cocop_runway_pavement/cocop_train_pavementmerged_to_pierwharf.txt'), DatasetClassUpdateRecord('cocop','val' , 'sidewalk_pavement', 'runway', 'cocop_runway_pavement/cocop_val_pavementmerged_to_airport_runway.txt'), DatasetClassUpdateRecord('cocop','val' , 'sidewalk_pavement', 'sidewalk_pavement', 'cocop_runway_pavement/cocop_val_pavementmerged_to_pavementmerged.txt'), # Phase 3,cocop_runway_road # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocop', 'train', 'road', 'runway', 'cocop_runway_road/cocop_train_road_to_airportrunway.txt'), DatasetClassUpdateRecord('cocop', 'train', 'road', 'sidewalk_pavement','cocop_runway_road/cocop_train_road_to_pavementsidewalk.txt'), DatasetClassUpdateRecord('cocop', 'train', 'road', 'road', 'cocop_runway_road/cocop_train_road_to_road.txt'), DatasetClassUpdateRecord('cocop', 'train', 'road', 'unlabeled', 'cocop_runway_road/cocop_train_road_to_unlabel.txt'), DatasetClassUpdateRecord('cocop', 'val', 'road', 'runway', 'cocop_runway_road/cocop_val_road_to_airportrunway.txt'), DatasetClassUpdateRecord('cocop', 'val', 'road', 'sidewalk_pavement','cocop_runway_road/cocop_val_road_to_pavementsidewalk.txt'), DatasetClassUpdateRecord('cocop', 'val', 'road', 'road', 'cocop_runway_road/cocop_val_road_to_road.txt'), # Phase 3,cocop_cocop_platform # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocop', 'train', 'platform', 'bridge', 'cocop_platform/cocop_train_cocop_platform_to_bridge.txt'), DatasetClassUpdateRecord('cocop', 'train', 'platform', 'unlabeled', 'cocop_platform/cocop_train_cocop_platform_to_None_of_these.txt'), DatasetClassUpdateRecord('cocop', 'train', 'platform', 'pier_wharf', 'cocop_platform/cocop_train_cocop_platform_to_pier-wharf.txt'), DatasetClassUpdateRecord('cocop', 'train', 'platform', 'platform', 'cocop_platform/cocop_train_cocop_platform_to_platform.txt'), DatasetClassUpdateRecord('cocop', 'val', 'platform', 'building', 'cocop_platform/cocop_val_cocop_platform_to_grandstand.txt'), DatasetClassUpdateRecord('cocop', 'val', 'platform', 'unlabeled', 'cocop_platform/cocop_val_cocop_platform_to_None_of_these.txt'), DatasetClassUpdateRecord('cocop', 'val', 'platform', 'pier_wharf', 'cocop_platform/cocop_val_cocop_platform_to_pier-wharf.txt'), DatasetClassUpdateRecord('cocop', 'val', 'platform', 'platform', 'cocop_platform/cocop_val_cocop_platform_to_platform.txt'), DatasetClassUpdateRecord('cocop', 'val', 'platform', 'playingfield', 'cocop_platform/cocop_val_cocop_platform_to_playingfield.txt'), DatasetClassUpdateRecord('cocop', 'val', 'platform', 'stage', 'cocop_platform/cocop_val_cocop_platform_to_stage.txt'), # Phase 3, cocop_cocop_bridge # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'bridge', 'cocop_bridge/cocop_train_cocop_bridge_to_bridge.txt'), DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'building', 'cocop_bridge/cocop_train_cocop_bridge_to_building.txt'), DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'building', 'cocop_bridge/cocop_train_cocop_bridge_to_grandstand.txt'), DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'building', 'cocop_bridge/cocop_train_cocop_bridge_to_house.txt'), DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'unlabeled', 'cocop_bridge/cocop_train_cocop_bridge_to_None_of_these.txt'), DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'pier_wharf', 'cocop_bridge/cocop_train_cocop_bridge_to_pier-wharf.txt'), DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'platform', 'cocop_bridge/cocop_train_cocop_bridge_to_platform.txt'), DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'runway', 'cocop_bridge/cocop_train_cocop_bridge_to_runway.txt'), DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'vegetation', 'cocop_bridge/cocop_train_cocop_bridge_to_vegetation.txt'), DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'bridge', 'cocop_bridge/cocop_val_cocop_bridge_to_bridge.txt'), DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'building', 'cocop_bridge/cocop_val_cocop_bridge_to_building.txt'), DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'building', 'cocop_bridge/cocop_val_cocop_bridge_to_house.txt'), DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'unlabeled', 'cocop_bridge/cocop_val_cocop_bridge_to_None_of_these.txt'), DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'pier_wharf', 'cocop_bridge/cocop_val_cocop_bridge_to_pier-wharf.txt'), # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocop', 'train', 'tent', 'awning', 'cocop_tent/cocop_train_cocop_tent_to_awning.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'base', 'cocop_tent/cocop_train_cocop_tent_to_base.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'building', 'cocop_tent/cocop_train_cocop_tent_to_building.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'fence', 'cocop_tent/cocop_train_cocop_tent_to_fence.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'unlabeled', 'cocop_tent/cocop_train_cocop_tent_to_None_of_these.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'billboard', 'cocop_tent/cocop_train_cocop_tent_to_signboard.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'tent', 'cocop_tent/cocop_train_cocop_tent_to_tent.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'umbrella', 'cocop_tent/cocop_train_cocop_tent_to_umbrella.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'vegetation', 'cocop_tent/cocop_train_cocop_tent_to_vegetation.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'wall', 'cocop_tent/cocop_train_cocop_tent_to_wall.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'window', 'cocop_tent/cocop_train_cocop_tent_to_window.txt'), DatasetClassUpdateRecord('cocop', 'train', 'tent', 'kite', 'cocop_tent/cocop_train_cocop_tent_to_windsurfingkite.txt'), # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cocop', 'val', 'tent', 'awning', 'cocop_tent/cocop_val_cocop_tent_to_awning.txt'), DatasetClassUpdateRecord('cocop', 'val', 'tent', 'building', 'cocop_tent/cocop_val_cocop_tent_to_building.txt'), DatasetClassUpdateRecord('cocop', 'val', 'tent', 'unlabeled', 'cocop_tent/cocop_val_cocop_tent_to_None_of_these.txt'), DatasetClassUpdateRecord('cocop', 'val', 'tent', 'tent', 'cocop_tent/cocop_val_cocop_tent_to_tent.txt'), DatasetClassUpdateRecord('cocop', 'val', 'tent', 'umbrella', 'cocop_tent/cocop_val_cocop_tent_to_umbrella.txt'), DatasetClassUpdateRecord('cocop', 'val', 'tent', 'window', 'cocop_tent/cocop_val_cocop_tent_to_window.txt') ] """ ADE20K """ ade20k_update_records = [ # Phase1 - fence guardrail, person-rider, shower-curtain, curtain-other, snow-mountain-hill-ade20k # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('ade20k-v1', 'train', 'fence', 'fence', 'ade20k_fence_guardrail/ade20k_fence_train.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'fence', 'fence', 'ade20k_fence_guardrail/ade20k_fence_val.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'fence', 'guard_rail', 'ade20k_fence_guardrail/ade20k_guardrail_train.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'fence', 'guard_rail', 'ade20k_fence_guardrail/ade20k_guardrail_val.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'curtain_other', 'curtain_other', 'ade20k_curtain/2019_10_31_ade20k_other-curtain_valsplit.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'curtain_other', 'shower_curtain', 'ade20k_curtain/2019_10_31_ade20k_showercurtain_valsplit.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'curtain_other', 'curtain_other', 'ade20k_curtain/2019_11_1_ade20k_other-curtain_trainsplit.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'curtain_other', 'shower_curtain', 'ade20k_curtain/2019_11_1_ade20k_showercurtain_trainsplit.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'mountain_hill', 'mountain_hill', 'ade20k_snow-mountain-hill/ade20k_mountainhill_train.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'mountain_hill', 'mountain_hill', 'ade20k_snow-mountain-hill/ade20k_mountainhill_val.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'mountain_hill', 'snow', 'ade20k_snow-mountain-hill/ade20k_snow_train.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'mountain_hill', 'snow', 'ade20k_snow-mountain-hill/ade20k_snow_val.txt'), # ADE20K_ade20k_animal # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'animal_other', 'ade20k_animal/ADE20K_train_ade20k_animal_to_animal-other.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'bear', 'ade20k_animal/ADE20K_train_ade20k_animal_to_bear.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'bird', 'ade20k_animal/ADE20K_train_ade20k_animal_to_bird.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'cat', 'ade20k_animal/ADE20K_train_ade20k_animal_to_cat.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'cow', 'ade20k_animal/ADE20K_train_ade20k_animal_to_cow.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'dog', 'ade20k_animal/ADE20K_train_ade20k_animal_to_dog.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'elephant', 'ade20k_animal/ADE20K_train_ade20k_animal_to_elephant.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'giraffe', 'ade20k_animal/ADE20K_train_ade20k_animal_to_giraffe.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'horse', 'ade20k_animal/ADE20K_train_ade20k_animal_to_horse.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'unlabeled', 'ade20k_animal/ADE20K_train_ade20k_animal_to_None_of_these.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other','plaything_other','ade20k_animal/ADE20K_train_ade20k_animal_to_plaything.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'sheep', 'ade20k_animal/ADE20K_train_ade20k_animal_to_sheep.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'zebra', 'ade20k_animal/ADE20K_train_ade20k_animal_to_zebra.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'animal_other', 'ade20k_animal/ADE20K_val_ade20k_animal_to_animal-other.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'bear', 'ade20k_animal/ADE20K_val_ade20k_animal_to_bear.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'bird', 'ade20k_animal/ADE20K_val_ade20k_animal_to_bird.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'cow', 'ade20k_animal/ADE20K_val_ade20k_animal_to_cow.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'elephant', 'ade20k_animal/ADE20K_val_ade20k_animal_to_elephant.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'horse', 'ade20k_animal/ADE20K_val_ade20k_animal_to_horse.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'sheep', 'ade20k_animal/ADE20K_val_ade20k_animal_to_sheep.txt'), # ade20k_ade20k_chest_of_drawers DatasetClassUpdateRecord('ade20k-v1', 'train', 'chest_of_drawers', 'cabinet', 'ade20k_chest_of_drawers/ade20k_train_ade20k_chest_of_drawers_to_cabinet.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'chest_of_drawers', 'chest_of_drawers', 'ade20k_chest_of_drawers/ade20k_train_ade20k_chest_of_drawers_to_chest-of-drawers.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'chest_of_drawers', 'nightstand', 'ade20k_chest_of_drawers/ade20k_train_ade20k_chest_of_drawers_to_nightstand.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'chest_of_drawers', 'unlabeled', 'ade20k_chest_of_drawers/ade20k_train_ade20k_chest_of_drawers_to_None_of_these.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'chest_of_drawers', 'cabinet', 'ade20k_chest_of_drawers/ade20k_val_ade20k_chest_of_drawers_to_cabinet.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'chest_of_drawers', 'chest_of_drawers', 'ade20k_chest_of_drawers/ade20k_val_ade20k_chest_of_drawers_to_chest-of-drawers.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'chest_of_drawers', 'desk', 'ade20k_chest_of_drawers/ade20k_val_ade20k_chest_of_drawers_to_desk.txt'), # TODO # ADE20K_ade20k_food DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'cake', 'ade20k_food/ADE20K_train_ade20k_food_to_cake.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'food_other', 'ade20k_food/ADE20K_train_ade20k_food_to_foodother.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'fruit_other', 'ade20k_food/ADE20K_train_ade20k_food_to_fruitother.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'pizza', 'ade20k_food/ADE20K_train_ade20k_food_to_pizza.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'sandwich', 'ade20k_food/ADE20K_train_ade20k_food_to_sandwich.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'unlabeled', 'ade20k_food/ADE20K_train_ade20k_food_to_unlabel.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'vegetation', 'ade20k_food/ADE20K_train_ade20k_food_to_vegetation.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'food_other', 'broccoli', 'ade20k_food/ADE20K_val_ade20k_food_to_broccoli.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'food_other', 'food_other', 'ade20k_food/ADE20K_val_ade20k_food_to_food-other.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'food_other', 'fruit_other', 'ade20k_food/ADE20K_val_ade20k_food_to_fruit-other.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'food_other', 'pizza', 'ade20k_food/ADE20K_val_ade20k_food_to_pizza.txt'), # ADE20K_ade20k_glass (ADE20K GLASS -> UNIVERSAL WINE-GLASS) DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'bottle', 'ade20k_glass/ADE20K_train_ade20k_glass_to_bottle.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'cup', 'ade20k_glass/ADE20K_train_ade20k_glass_to_cup.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'mirror', 'ade20k_glass/ADE20K_train_ade20k_glass_to_mirror.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'unlabeled', 'ade20k_glass/ADE20K_train_ade20k_glass_to_None_of_these.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'window', 'ade20k_glass/ADE20K_train_ade20k_glass_to_window.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'wine_glass', 'ade20k_glass/ADE20K_train_ade20k_glass_to_wine_glass.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'wine_glass', 'cup', 'ade20k_glass/ADE20K_val_ade20k_glass_to_cup.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'wine_glass', 'unlabeled', 'ade20k_glass/ADE20K_val_ade20k_glass_to_unlabel.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'wine_glass', 'window', 'ade20k_glass/ADE20K_val_ade20k_glass_to_window.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'wine_glass', 'wine_glass', 'ade20k_glass/ADE20K_val_ade20k_glass_to_wine_glass.txt'), # ADE20K_ade20k_motorcyclist_bicyclist (ADE20K PERSON -> UNIVERSAL PERSON-NON-RIDER) DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'bicyclist', 'ade20k_motorcyclist_bicyclist/ADE20K_train_ade20k_motorcyclist_bicyclist_to_bicyclist.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'motorcyclist', 'ade20k_motorcyclist_bicyclist/ADE20K_train_ade20k_motorcyclist_bicyclist_to_motorcyclist.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'rider_other', 'ade20k_motorcyclist_bicyclist/ADE20K_train_ade20k_motorcyclist_bicyclist_to_rider-other.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'person', 'bicyclist', 'ade20k_motorcyclist_bicyclist/ADE20K_val_ade20k_motorcyclist_bicyclist_to_bicyclist.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'person', 'motorcyclist', 'ade20k_motorcyclist_bicyclist/ADE20K_val_ade20k_motorcyclist_bicyclist_to_motorcyclist.txt'), # ade20k_ade20k_nonrider (ADE20K PERSON -> UNIVERSAL PERSON-NON-RIDER) DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'bicyclist', 'ade20k_nonrider/ade20k_train_ade20k_nonrider_to_bicyclist.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'person', 'ade20k_nonrider/ade20k_train_ade20k_nonrider_to_person-non-rider.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'person', 'person', 'ade20k_nonrider/ade20k_val_ade20k_nonrider_to_person-non-rider.txt'), #ade20k_ade20k_table (ADE20K TABLE -> UNIVERSAL TABLE) DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'bathroom_counter', 'ade20k_table/ade20k_train_ade20k_table_to_bathroom-counter.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'cabinet', 'ade20k_table/ade20k_train_ade20k_table_to_cabinet.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'counter_other', 'ade20k_table/ade20k_train_ade20k_table_to_counter-other.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'desk', 'ade20k_table/ade20k_train_ade20k_table_to_desk-and-table.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'desk', 'ade20k_table/ade20k_train_ade20k_table_to_desk.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'kitchen_island', 'ade20k_table/ade20k_train_ade20k_table_to_kitchen-island.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'nightstand', 'ade20k_table/ade20k_train_ade20k_table_to_nightstand.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'unlabeled', 'ade20k_table/ade20k_train_ade20k_table_to_None_of_these.txt'), DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'table', 'ade20k_table/ade20k_train_ade20k_table_to_table.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'counter_other', 'ade20k_table/ade20k_val_ade20k_table_to_counter-other.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'desk', 'ade20k_table/ade20k_val_ade20k_table_to_desk-and-table.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'desk', 'ade20k_table/ade20k_val_ade20k_table_to_desk.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'kitchen_island', 'ade20k_table/ade20k_val_ade20k_table_to_kitchen-island.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'nightstand', 'ade20k_table/ade20k_val_ade20k_table_to_nightstand.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'unlabeled', 'ade20k_table/ade20k_val_ade20k_table_to_None_of_these.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'stool', 'ade20k_table/ade20k_val_ade20k_table_to_stool.txt'), DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'table', 'ade20k_table/ade20k_val_ade20k_table_to_table.txt'), # Phase 3, ade20k_plaything # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('ade20k', 'train', 'plaything_other', 'plaything_other','ade20k_plaything/ade20k_train_ade20k_plaything_to_plaything-other.txt'), DatasetClassUpdateRecord('ade20k', 'train', 'plaything_other', 'teddy_bear', 'ade20k_plaything/ade20k_train_ade20k_plaything_to_teddy-bear.txt'), DatasetClassUpdateRecord('ade20k', 'val', 'plaything_other', 'plaything_other', 'ade20k_plaything/ade20k_val_ade20k_plaything_to_plaything-other.txt'), DatasetClassUpdateRecord('ade20k', 'val', 'plaything_other', 'teddy_bear', 'ade20k_plaything/ade20k_val_ade20k_plaything_to_teddy-bear.txt') ] cityscapes_update_records = [ # cityscapes_rider, Phase 2 (CITYSCAPES RIDER -> UNVERISAL RIDER-OTHER) # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'bicycle', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_bicycle.txt'), DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'bicyclist', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_bicyclist.txt'), DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'motorcyclist', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_motorcyclist.txt'), DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'unlabeled', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_None_of_these.txt'), DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'person', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_person-non-rider.txt'), DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'rider_other', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_rider-other.txt'), DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'bicycle', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_bicycle.txt'), DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'bicyclist', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_bicyclist.txt'), DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'motorcyclist', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_motorcyclist.txt'), DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'unlabeled', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_None_of_these.txt'), DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'person', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_person-non-rider.txt'), DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'rider_other', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_rider-other.txt') ] bdd_update_records = [ # bdd person, Phase 2 (BDD PERSON -> UNIVERSAL PERSON-NONRIDER) # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('bdd', 'train', 'person', 'bicyclist', 'bdd_person/bdd_train_bdd_person_to_bicyclist.txt'), DatasetClassUpdateRecord('bdd', 'train', 'person', 'motorcyclist', 'bdd_person/bdd_train_bdd_person_to_motorcyclist.txt'), DatasetClassUpdateRecord('bdd', 'train', 'person', 'unlabeled', 'bdd_person/bdd_train_bdd_person_to_None_of_these.txt'), DatasetClassUpdateRecord('bdd', 'train', 'person', 'person', 'bdd_person/bdd_train_bdd_person_to_person-non-rider.txt'), DatasetClassUpdateRecord('bdd', 'val', 'person', 'bicyclist', 'bdd_person/bdd_val_bdd_person_to_bicyclist.txt'), DatasetClassUpdateRecord('bdd', 'val', 'person', 'person', 'bdd_person/bdd_val_bdd_person_to_person-non-rider.txt'), DatasetClassUpdateRecord('bdd', 'val', 'person', 'unlabeled', 'bdd_person/bdd_val_bdd_person_to_unlabel.txt'), # bdd_rider, Phase 2 (BDD RIDER -> UNIVERSAL RIDER-OTHER)) # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'bicyclist', 'bdd_rider/bdd_train_bdd_rider_to_bicyclist.txt'), DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'motorcyclist', 'bdd_rider/bdd_train_bdd_rider_to_motorcyclist.txt'), DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'unlabeled', 'bdd_rider/bdd_train_bdd_rider_to_None_of_these.txt'), DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'person', 'bdd_rider/bdd_train_bdd_rider_to_person-non-rider.txt'), DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'rider_other', 'bdd_rider/bdd_train_bdd_rider_to_rider-other.txt'), DatasetClassUpdateRecord('bdd', 'val', 'rider_other', 'bicyclist', 'bdd_rider/bdd_val_bdd_rider_to_bicyclist.txt'), DatasetClassUpdateRecord('bdd', 'val', 'rider_other', 'motorcyclist', 'bdd_rider/bdd_val_bdd_rider_to_motorcyclist.txt'), DatasetClassUpdateRecord('bdd', 'val', 'rider_other', 'person', 'bdd_rider/bdd_val_bdd_rider_to_person-non-rider.txt'), DatasetClassUpdateRecord('bdd', 'val', 'rider_other', 'rider_other', 'bdd_rider/bdd_val_bdd_rider_to_rider-other.txt') ] idd_update_records = [ # idd_idd_rider, Phase 2 (IDD-NEW RIDER -> UNIVERSAL RIDER-OTHER) # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'backpack', 'idd_rider/idd_train_idd_rider_to_backpack.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'bag', 'idd_rider/idd_train_idd_rider_to_bag.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'bicycle', 'idd_rider/idd_train_idd_rider_to_bicycle.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'bicyclist', 'idd_rider/idd_train_idd_rider_to_bicyclist.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'box', 'idd_rider/idd_train_idd_rider_to_box.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'motorcycle', 'idd_rider/idd_train_idd_rider_to_motorcycle.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'motorcyclist', 'idd_rider/idd_train_idd_rider_to_motorcyclist.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'unlabeled', 'idd_rider/idd_train_idd_rider_to_None_of_these.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'person', 'idd_rider/idd_train_idd_rider_to_person-non-rider.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'pole', 'idd_rider/idd_train_idd_rider_to_pole.txt'), DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'rider_other', 'idd_rider/idd_train_idd_rider_to_rider-other.txt'), DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'bicyclist', 'idd_rider/idd_val_idd_rider_to_bicyclist.txt'), DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'motorcyclist', 'idd_rider/idd_val_idd_rider_to_motorcyclist.txt'), DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'unlabeled', 'idd_rider/idd_val_idd_rider_to_None_of_these.txt'), DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'person', 'idd_rider/idd_val_idd_rider_to_person-non-rider.txt'), DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'rider_other', 'idd_rider/idd_val_idd_rider_to_rider-other.txt') ] # Note: we refer to the `test` split as `val`, since `val` not easily defined/found. sunrgbd_update_records = [ # sunrgbd_chair # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'armchair', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_armchair.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'bench', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_bench.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'chair_other', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_chairother.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'seat', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_seat.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'couch', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_sofa.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'stool', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_stool.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'swivel_chair', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_swivelchair.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'unlabeled', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_unlabel.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'armchair', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_armchair.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'bench', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_bench.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'cabinet', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_cabinet.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'chair_other', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_chair-other.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'door', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_door.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'unlabeled', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_None_of_these.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'ottoman', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_ottoman.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'seat', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_seat.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'couch', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_sofa.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'stool', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_stool.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'swivel_chair', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_swivel-chair.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'table', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_table.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'wall', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_wall.txt'), # sunrgbd_counter DatasetClassUpdateRecord('sunrgbd', 'val', 'counter_other', 'bathroom_counter','sunrgbd_counter/sunrgbd_test_sunrgbd_counter_to_bathroomcounter.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'counter_other', 'counter_other', 'sunrgbd_counter/sunrgbd_test_sunrgbd_counter_to_counter-other.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'counter_other', 'unlabeled', 'sunrgbd_counter/sunrgbd_test_sunrgbd_counter_to_unlabel.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'bathroom_counter','sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_bathroom-counter.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'cabinet', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_cabinetmerged.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'counter_other', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_counter-other.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'desk', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_desk.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'kitchen_island', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_kitchen-island.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'unlabeled', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_None_of_these.txt'), # sunrgbd_lamp (TODO: map it to "lamp" inside of the universal tax.) DatasetClassUpdateRecord('sunrgbd', 'val', 'light_other', 'lamp', 'sunrgbd_lamp/sunrgbd_test_sunrgbd_lamp_to_lamp.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'light_other', 'sconce', 'sunrgbd_lamp/sunrgbd_test_sunrgbd_lamp_to_sconce.txt'), DatasetClassUpdateRecord('sunrgbd', 'val', 'light_other', 'unlabeled', 'sunrgbd_lamp/sunrgbd_test_sunrgbd_lamp_to_unlabel.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'chandelier', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_chandelier.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'lamp', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_lamp.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'light_other', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_light-other.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'unlabeled', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_None_of_these.txt'), DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'sconce', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_sconce.txt') ] mapillary_update_records = [ # mapillarypublic_water # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'fountain', 'mapillarypublic_water/mapillarypublic_train_water_to_fountain.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'unlabeled', 'mapillarypublic_water/mapillarypublic_train_water_to_None_of_these.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'river_lake', 'mapillarypublic_water/mapillarypublic_train_water_to_river-lake.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'sea', 'mapillarypublic_water/mapillarypublic_train_water_to_sea.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'water_other','mapillarypublic_water/mapillarypublic_train_water_to_water-other.txt'), DatasetClassUpdateRecord('mapillary-public65', 'val', 'water_other', 'fountain', 'mapillarypublic_water/mapillarypublic_val_water_to_fountain.txt'), DatasetClassUpdateRecord('mapillary-public65', 'val', 'water_other', 'river_lake', 'mapillarypublic_water/mapillarypublic_val_water_to_river-lake.txt'), DatasetClassUpdateRecord('mapillary-public65', 'val', 'water_other', 'sea', 'mapillarypublic_water/mapillarypublic_val_water_to_sea.txt'), DatasetClassUpdateRecord('mapillary-public65', 'val', 'water_other', 'water_other', 'mapillarypublic_water/mapillarypublic_val_water_to_water-other.txt'), # mapillarypublic_groundanimal # dataset # split # orig u name # new u name # path to txt file DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'animal_other', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_animalother.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'bird', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_bird.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'cat', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_cat.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'cow', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_cow.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'dog', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_dog.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'horse', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_horse.txt'), DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'unlabeled', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_none.txt'), DatasetClassUpdateRecord('mapillary-public65', 'val', 'dog', 'bird', 'mapillarypublic_groundanimal/mapillaryacademic_val_groundanimal_to_bird.txt'), DatasetClassUpdateRecord('mapillary-public65', 'val', 'dog', 'dog', 'mapillarypublic_groundanimal/mapillaryacademic_val_groundanimal_to_dog.txt'), DatasetClassUpdateRecord('mapillary-public65', 'val', 'dog', 'unlabeled', 'mapillarypublic_groundanimal/mapillaryacademic_val_groundanimal_to_none.txt') ]
105.911439
195
0.785433
6,608
57,404
6.407234
0.034201
0.221923
0.107773
0.088122
0.939394
0.897494
0.739885
0.420393
0.295236
0.194147
0
0.019679
0.076702
57,404
541
196
106.107209
0.779155
0.050763
0
0
0
0
0.638556
0.418052
0
0
0
0.001848
0
1
0
false
0
0.002525
0
0.002525
0
0
0
0
null
1
0
0
1
1
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
1
1
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
2d6b442d289e60626f3caad22d37bc5cb6431b4b
9,758
py
Python
kvlnode/core/KVLSegment.py
paoloriccardi/key-value-log
bc08d26ec614d9570eef5ca567dc69cc1f9a6ef7
[ "BSD-3-Clause" ]
null
null
null
kvlnode/core/KVLSegment.py
paoloriccardi/key-value-log
bc08d26ec614d9570eef5ca567dc69cc1f9a6ef7
[ "BSD-3-Clause" ]
null
null
null
kvlnode/core/KVLSegment.py
paoloriccardi/key-value-log
bc08d26ec614d9570eef5ca567dc69cc1f9a6ef7
[ "BSD-3-Clause" ]
null
null
null
import sys import os import json from datetime import datetime #Segment with JSON as values delimited by {} class KVLSegmentJSON(): def __init__(self,filename): self.homedir = os.path.join(os.path.abspath(os.path.dirname(__file__)), "files/") self.filename = filename self.filepath = self.homedir + self.filename try: self.file = open(self.filepath,"a+") except OSError: print ("Could not open file" + self.filepath + "\n") sys.exit() def appendKeyValue(self, key, value): if not self.checkValueFormat(value): print (value) print ("json format error in input") exit() #stringValue = json.dumps(value) elementString = str(key) + ":" + value + ";" rwpointer = self.file.tell() if rwpointer > 1: #takes into account element separator ; rwpointer = rwpointer try: self.file.write(elementString) except OSError: print ("Could not append Key:Value \n") return -1 return rwpointer def checkValueFormat(self,value): try: json.loads(value) except ValueError as e: print("Error in JSON format value " + str(e)) return False return True def retrieveElement(self,offset): self.file.seek(offset) elementString = "" while True: character = self.file.read(1) if not character: break if character == "}": elementString = elementString + character break elementString = elementString + character return elementString def retrieveValue(self,offset): self.file.seek(offset) valueString = "" scanningValue = False while True: character = self.file.read(1) if not character: break if not scanningValue and character == "{": valueString = character scanningValue = True elif not scanningValue and character != "{": pass elif scanningValue and character == "}": valueString = valueString + character break else: valueString = valueString + character return valueString def createIndex(self): #scan one char at the time the segment file from beginning to end which is very inefficient #this method should be used only when a consistent/updated version of index is not available (e.g. from an existing bucket) self.file.seek(0) index = {} prevchar = self.file.read(1) if not prevchar: return index #non trivial case self.file.seek(0) key = "" scanningKey = True prevchar = "" rwindex = 0 offset = 0 for char in self.file.read(): if scanningKey and char == ":": scanningKey = False index[key]=offset key = "" elif scanningKey and char != ":": key = key + char elif not scanningKey and prevchar == "}" and char == ";": scanningKey = True offset = rwindex +1 prevchar = char rwindex = rwindex + 1 return index def getTombstoneValue(self): return "{}" def flush(self): #close segment file self.file.close() def attachNewFile(self,newFilename): self.filename = newFilename try: self.filepath = self.homedir + self.filename self.file = open(self.filepath, "a+") except OSError: print ("Could not open file" + newFilename + "\n") sys.exit() def shrinkToNewFile(self): inmemoryKV = self.inMemoryKeyValue() newFilename = self.generateFilename() self.flush() self.attachNewFile(newFilename) for key,value in inmemoryKV.items(): self.appendKeyValue(key,value) newIndex = self.createIndex() return newIndex def inMemoryKeyValue(self): KVDict = {} index = self.createIndex() for key,offset in index.items(): value = self.retrieveValue(offset) KVDict[key]=value return KVDict def initializeSegment(self,kvdict): #careful when using this method straight from the API initialize, this method expects a dictionary #in case of nested json in request from the API the situation should be handled before calling #appendKeyValue (where value should be a valid JSON) newFilename = self.generateFilename() self.file.flush() self.attachNewFile(newFilename) for key,value in kvdict.items(): self.appendKeyValue(key,value) newIndex = self.createIndex() return newIndex def generateFilename(self): now = datetime.now() prefix = now.strftime('%f') newFilename = prefix + self.filename return newFilename #Segment with value equal to simple value, key:values separated by ; class KVLSegmentSimpleValue(): def __init__(self,filename): self.homedir = os.path.join(os.path.abspath(os.path.dirname(__file__)), "files/") self.filename = filename self.filepath = self.homedir + self.filename try: self.file = open(self.filepath,"a+") except OSError: print ("Could not open file" + self.filepath + "\n") sys.exit() def appendKeyValue(self, key, value): elementString = str(key) + ":" + str(value) + ";" rwpointer = self.file.tell() if rwpointer > 1: #takes into account element separator ; rwpointer = rwpointer try: self.file.write(elementString) except OSError: print ("Could not append Key:Value \n") return -1 return rwpointer def retrieveElement(self,offset): self.file.seek(offset) elementString = "" while True: character = self.file.read(1) if not character: break if character == ";": elementString = elementString break elementString = elementString + character return elementString def retrieveValue(self,offset): self.file.seek(offset) valueString = "" scanningValue = False while True: character = self.file.read(1) if not character: break if not scanningValue and character == ":": scanningValue = True elif not scanningValue and character != ":": pass elif scanningValue and character == ";": valueString = valueString break else: valueString = valueString + character return valueString def createIndex(self): #scan one char at the time the segment file from beginning to end which is very inefficient #this method should be used only when a consistent/updated version of index is not available (e.g. from an existing bucket) self.file.seek(0) index = {} if not self.file.read(1): return index #non trivial case self.file.seek(0) key = "" scanningKey = True rwindex = 0 offset = 0 for char in self.file.read(): if scanningKey and char == ":": scanningKey = False index[key]=offset key = "" elif scanningKey and char != ":": key = key + char elif not scanningKey and char != ";": pass elif not scanningKey and char == ";": scanningKey = True offset = rwindex +1 rwindex = rwindex + 1 return index def getTombstoneValue(self): return "" def checkValueFormat(self,value): return True def flush(self): #close segment file self.file.close() def attachNewFile(self,newFilename): self.filename = newFilename try: self.filepath = self.homedir + self.filename self.file = open(self.filepath, "a+") except OSError: print ("Could not open file" + newFilename + "\n") sys.exit() def shrinkToNewFile(self): inmemoryKV = self.inMemoryKeyValue() newFilename = self.generateFilename() self.flush() self.attachNewFile(newFilename) for key,value in inmemoryKV.items(): self.appendKeyValue(key,value) newIndex = self.createIndex() return newIndex def inMemoryKeyValue(self): KVDict = {} index = self.createIndex() for key,offset in index.items(): value = self.retrieveValue(offset) KVDict[key]=value return KVDict def initializeSegment(self,kvdict): newFilename = self.generateFilename() self.file.flush() self.attachNewFile(newFilename) for key,value in kvdict.items(): self.appendKeyValue(key,value) newIndex = self.createIndex() return newIndex def generateFilename(self): now = datetime.now() prefix = now.strftime('%f') newFilename = prefix + self.filename return newFilename
31.993443
131
0.552572
958
9,758
5.611691
0.160752
0.041667
0.017857
0.02567
0.857329
0.853237
0.849888
0.836496
0.836496
0.836496
0
0.003535
0.362267
9,758
304
132
32.098684
0.860357
0.097356
0
0.85259
0
0
0.027411
0
0
0
0
0
0
1
0.103586
false
0.011952
0.015936
0.011952
0.227092
0.035857
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
2d9f81ee38a3d99745cc8cb0159333b13ca2f76e
6,750
py
Python
pirates/leveleditor/worldData/shipNavyMerchant3.py
Willy5s/Pirates-Online-Rewritten
7434cf98d9b7c837d57c181e5dabd02ddf98acb7
[ "BSD-3-Clause" ]
81
2018-04-08T18:14:24.000Z
2022-01-11T07:22:15.000Z
pirates/leveleditor/worldData/shipNavyMerchant3.py
Willy5s/Pirates-Online-Rewritten
7434cf98d9b7c837d57c181e5dabd02ddf98acb7
[ "BSD-3-Clause" ]
4
2018-09-13T20:41:22.000Z
2022-01-08T06:57:00.000Z
pirates/leveleditor/worldData/shipNavyMerchant3.py
Willy5s/Pirates-Online-Rewritten
7434cf98d9b7c837d57c181e5dabd02ddf98acb7
[ "BSD-3-Clause" ]
26
2018-05-26T12:49:27.000Z
2021-09-11T09:11:59.000Z
from pandac.PandaModules import Point3, VBase3, Vec4 objectStruct = {'Objects': {'1189040144.92gjeon': {'Type': 'Ship Part','Name': 'shipNavyMerchant3','Category': '13: War Galleon','File': '','Flagship': True,'Objects': {'1189040455.83gjeon': {'Type': 'Spawn Node','Aggro Radius': '12.0000','AnimSet': 'default','Hpr': Point3(0.0, 0.0, 0.0),'Min Population': '1','Patrol Radius': '12.0000','Pause Chance': '100','Pause Duration': '30','Pos': Point3(0.957, 13.753, 52.555),'Scale': VBase3(1.0, 1.0, 1.0),'Spawnables': 'Area','Start State': 'Patrol','Team': 'default','Visual': {'Color': (0, 0, 0.65, 1),'Model': 'models/misc/smiley'}},'1189040561.39gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-26.495, 45.455, 52.564),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040572.72gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-22.419, -3.694, 52.555),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040581.14gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(26.193, 46.064, 52.564),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040584.44gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(21.954, -5.448, 52.555),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040632.77gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-27.021, 84.795, 76.512),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040642.09gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-34.475, 115.512, 72.838),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040647.58gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(35.397, 116.217, 72.754),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040650.8gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': 100,'Pause Duration': 30,'Pos': Point3(27.643, 86.15, 76.35),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040698.55gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': 100,'Pause Duration': 30,'Pos': Point3(-21.446, -63.181, 94.129),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040703.08gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': 100,'Pause Duration': 30,'Pos': Point3(20.874, -63.84, 94.213),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040766.88gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-1.089, 51.312, 52.565),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040838.08gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(0.676, 96.848, 75.059),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189041023.56gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-0.07, -10.163, 53.999),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}}},'Respawns': True,'Team': 'EvilNavy','Visual': {'Model': ['models/shipparts/merchantL3-geometry_High', 'models/shipparts/merchantL3-collisions', 'models/shipparts/merchantCabinAL3-collisions', 'models/shipparts/merchantCabinAL3-geometry_High']}}},'Node Links': [['1189040766.88gjeon', '1189040561.39gjeon', 'Bi-directional'], ['1189040766.88gjeon', '1189040581.14gjeon', 'Bi-directional'], ['1189040561.39gjeon', '1189040632.77gjeon', 'Bi-directional'], ['1189040642.09gjeon', '1189040632.77gjeon', 'Bi-directional'], ['1189040642.09gjeon', '1189040647.58gjeon', 'Bi-directional'], ['1189040650.8gjeon', '1189040647.58gjeon', 'Bi-directional'], ['1189040581.14gjeon', '1189040650.8gjeon', 'Bi-directional'], ['1189040581.14gjeon', '1189040647.58gjeon', 'Bi-directional'], ['1189040838.08gjeon', '1189040632.77gjeon', 'Bi-directional'], ['1189040650.8gjeon', '1189040838.08gjeon', 'Bi-directional'], ['1189040561.39gjeon', '1189040455.83gjeon', 'Bi-directional'], ['1189040581.14gjeon', '1189040455.83gjeon', 'Bi-directional'], ['1189040561.39gjeon', '1189040572.72gjeon', 'Bi-directional'], ['1189040581.14gjeon', '1189040584.44gjeon', 'Bi-directional'], ['1189040572.72gjeon', '1189040698.55gjeon', 'Bi-directional'], ['1189040703.08gjeon', '1189040584.44gjeon', 'Bi-directional'], ['1189040584.44gjeon', '1189041023.56gjeon', 'Bi-directional'], ['1189040572.72gjeon', '1189041023.56gjeon', 'Bi-directional']],'Layers': {},'ObjectIds': {'1189040144.92gjeon': '["Objects"]["1189040144.92gjeon"]','1189040455.83gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040455.83gjeon"]','1189040561.39gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040561.39gjeon"]','1189040572.72gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040572.72gjeon"]','1189040581.14gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040581.14gjeon"]','1189040584.44gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040584.44gjeon"]','1189040632.77gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040632.77gjeon"]','1189040642.09gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040642.09gjeon"]','1189040647.58gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040647.58gjeon"]','1189040650.8gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040650.8gjeon"]','1189040698.55gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040698.55gjeon"]','1189040703.08gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040703.08gjeon"]','1189040766.88gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040766.88gjeon"]','1189040838.08gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040838.08gjeon"]','1189041023.56gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189041023.56gjeon"]'}}
3,375
6,697
0.657926
941
6,750
4.717322
0.167906
0.038297
0.038522
0.037846
0.434332
0.412255
0.391079
0.391079
0.3877
0.378914
0
0.272067
0.061778
6,750
2
6,697
3,375
0.428865
0
0
0
0
0
0.606429
0.16679
0
0
0
0
0
1
0
false
0
0.5
0
0.5
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
1
0
0
1
1
1
1
0
0
0
0
0
1
0
null
0
0
0
0
0
0
0
0
1
0
0
0
0
7
933663beb3b646af38f4e483e5c865f8634c1bfb
16,451
py
Python
tests/test__calculation.py
SteamPeKa/krippendorffs_alpha
c3d4f3eacaf418aeb22d30759594ca567dce9ecb
[ "MIT" ]
1
2020-10-28T09:37:13.000Z
2020-10-28T09:37:13.000Z
tests/test__calculation.py
SteamPeKa/krippendorffs_alpha
c3d4f3eacaf418aeb22d30759594ca567dce9ecb
[ "MIT" ]
null
null
null
tests/test__calculation.py
SteamPeKa/krippendorffs_alpha
c3d4f3eacaf418aeb22d30759594ca567dce9ecb
[ "MIT" ]
null
null
null
# coding=utf-8 # Creation date: 27 окт. 2020 # Creation time: 18:52 # Creator: SteamPeKa import csv import json import os import numpy import pytest import krippendorffs_alpha import testing_utils # Example E data matrix OBSERVER_A_DATA = numpy.array([ [1, 0, 0, 0, 0, 1, 0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0], [0, 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ]).T OBSERVER_B_DATA = numpy.array([ [1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0], [0, 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 1], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0], ]).T OBSERVER_C_DATA = numpy.array([ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0], [0, 1, 1, 1, 0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0], ]).T OBSERVER_D_DATA = numpy.array([ [1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 1, 0], [0, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0], [0, 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0], ]).T DATA_MATRIX = numpy.zeros((4, 12, 5)) DATA_MATRIX[0, :, :] = OBSERVER_A_DATA DATA_MATRIX[1, :, :] = OBSERVER_B_DATA DATA_MATRIX[2, :, :] = OBSERVER_C_DATA DATA_MATRIX[3, :, :] = OBSERVER_D_DATA DATA_MATRIX.flags.writeable = False OBSERVER_A_DATA.flags.writeable = False OBSERVER_B_DATA.flags.writeable = False OBSERVER_C_DATA.flags.writeable = False OBSERVER_D_DATA.flags.writeable = False class TestMakeCoincidencesMatrixFromDataMatrix(object): def test_data_from_example_no_omit(self): expected_value_by_unit_matrix = numpy.array([ [3, 0, 0, 0, 0, 1, 0, 3, 0, 0, 2, 0], [0, 3, 0, 0, 4, 1, 0, 1, 4, 0, 0, 0], [0, 1, 4, 4, 0, 1, 0, 0, 0, 0, 0, 1], [0, 0, 0, 0, 0, 1, 4, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 3, 0, 0], ]) actual_value_by_unit_matrix = krippendorffs_alpha._calculation._make_value_by_unit_matrix_from_data_matrix( DATA_MATRIX, omit_unpairable=False ) testing_utils.assert_equal_tensors(expected_value_by_unit_matrix, actual_value_by_unit_matrix) def test_data_from_example_omit_unpairable(self): expected_value_by_unit_matrix = numpy.array([ [3, 0, 0, 0, 0, 1, 0, 3, 0, 0, 2], [0, 3, 0, 0, 4, 1, 0, 1, 4, 0, 0], [0, 1, 4, 4, 0, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 1, 4, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 3, 0], ]) actual_value_by_unit_matrix = krippendorffs_alpha._calculation._make_value_by_unit_matrix_from_data_matrix( DATA_MATRIX, omit_unpairable=True ) testing_utils.assert_equal_tensors(expected_value_by_unit_matrix, actual_value_by_unit_matrix) # noinspection PyPep8Naming class Test_CalcAlpha(object): def test_e_nominal(self): metric_tensor = krippendorffs_alpha.metrics.NominalMetric().get_metric_tensor(list(range(1, 6)), symmetric=False) actual_alpha = krippendorffs_alpha._calculation._calc_alpha(DATA_MATRIX, metric_tensor) assert actual_alpha == pytest.approx(0.743, abs=0.001) def test_e_interval(self): metric_tensor = krippendorffs_alpha.metrics.IntervalMetric().get_metric_tensor(list(range(1, 6)), symmetric=False) actual_alpha = krippendorffs_alpha._calculation._calc_alpha(DATA_MATRIX, metric_tensor) assert actual_alpha == pytest.approx(0.849, abs=0.001) class TestCalcAlphaFromListOrLists(object): def test_e_nominal(self): with open(os.path.join("tests", "example_E_data.tsv"), "r") as f: input_table = csv.reader(f, delimiter="\t") prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists( input_table=input_table, header=True, row_legend=True, upper_level="observer", value_constructor=lambda s: int(s.strip()) if s.strip() != "NULL" else None ) testing_utils.assert_equal_tensors(DATA_MATRIX, prepared_data.answers_tensor) actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "nominal") assert actual_alpha == pytest.approx(0.743, abs=0.001) def test_e_interval(self): with open(os.path.join("tests", "example_E_data.tsv"), "r") as f: input_table = csv.reader(f, delimiter="\t") prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists( input_table=input_table, header=True, row_legend=True, upper_level="observer", value_constructor=lambda s: int(s.strip()) if s.strip() != "NULL" else None ) testing_utils.assert_equal_tensors(DATA_MATRIX, prepared_data.answers_tensor) actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "interval") assert actual_alpha == pytest.approx(0.849, abs=0.001) def test_wikipedia_nominal(self): with open(os.path.join("tests", "example_wikipedia.csv"), "r") as f: input_table = csv.reader(f, delimiter=",") prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists( input_table=input_table, header=True, row_legend=True, upper_level="observer", value_constructor=lambda s: int(s.strip()) if s.strip() != "*" else None ) actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "nominal") assert actual_alpha == pytest.approx(0.691, abs=0.001) def test_wikipedia_interval(self): with open(os.path.join("tests", "example_wikipedia.csv"), "r") as f: input_table = csv.reader(f, delimiter=",") prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists( input_table=input_table, header=True, row_legend=True, upper_level="observer", value_constructor=lambda s: int(s.strip()) if s.strip() != "*" else None ) actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "interval") assert actual_alpha == pytest.approx(0.811, abs=0.001) class TestCalcAlphaFromDictOfDicts(object): def test_e_nominal(self): with open(os.path.join("tests", "example_E_data.json"), "r") as f: input_table = json.load(f) prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts( input_table=input_table, upper_level="observer", value_constructor=int ) actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "nominal") assert actual_alpha == pytest.approx(0.743, abs=0.001) def test_e_interval(self): with open(os.path.join("tests", "example_E_data.json"), "r") as f: input_table = json.load(f) prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts( input_table=input_table, upper_level="observer", value_constructor=int ) actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "interval") assert actual_alpha == pytest.approx(0.849, abs=0.001) def test_wikipedia_nominal(self): with open(os.path.join("tests", "example_wikipedia.json"), "r") as f: input_table = json.load(f) prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts( input_table=input_table, upper_level="observer", value_constructor=int ) actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "nominal") assert actual_alpha == pytest.approx(0.691, abs=0.001) def test_wikipedia_interval(self): with open(os.path.join("tests", "example_wikipedia.json"), "r") as f: input_table = json.load(f) prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts( input_table=input_table, upper_level="observer", value_constructor=int ) actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "interval") assert actual_alpha == pytest.approx(0.811, abs=0.001) class Test_CalcAlphaByPrecomputesFromListOfLists(object): def test_e_nominal(self): metric_name = "nominal" with open(os.path.join("tests", "example_E_data.tsv"), "r") as f: input_table = csv.reader(f, delimiter="\t") prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists( input_table=input_table, header=True, row_legend=True, upper_level="observer", value_constructor=lambda s: int(s.strip()) if s.strip() != "NULL" else None ) (assignment_matrix, full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes( prepared_data=prepared_data, metric=metric_name ) actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes( assignment_matrix=assignment_matrix, full_cross_disagreement_tensor=full_cross_disagreement_tensor ) assert actual_alpha == pytest.approx(0.743, abs=0.001) def test_e_interval(self): metric_name = "interval" with open(os.path.join("tests", "example_E_data.tsv"), "r") as f: input_table = csv.reader(f, delimiter="\t") prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists( input_table=input_table, header=True, row_legend=True, upper_level="observer", value_constructor=lambda s: int(s.strip()) if s.strip() != "NULL" else None ) (assignment_matrix, full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes( prepared_data=prepared_data, metric=metric_name ) actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes( assignment_matrix=assignment_matrix, full_cross_disagreement_tensor=full_cross_disagreement_tensor ) assert actual_alpha == pytest.approx(0.849, abs=0.001) def test_wikipedia_nominal(self): metric_name = "nominal" with open(os.path.join("tests", "example_wikipedia.csv"), "r") as f: input_table = csv.reader(f, delimiter=",") prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists( input_table=input_table, header=True, row_legend=True, upper_level="observer", value_constructor=lambda s: int(s.strip()) if s.strip() != "*" else None ) (assignment_matrix, full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes( prepared_data=prepared_data, metric=metric_name ) actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes( assignment_matrix=assignment_matrix, full_cross_disagreement_tensor=full_cross_disagreement_tensor ) assert actual_alpha == pytest.approx(0.691, abs=0.001) def test_wikipedia_interval(self): metric_name = "interval" with open(os.path.join("tests", "example_wikipedia.csv"), "r") as f: input_table = csv.reader(f, delimiter=",") prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists( input_table=input_table, header=True, row_legend=True, upper_level="observer", value_constructor=lambda s: int(s.strip()) if s.strip() != "*" else None ) (assignment_matrix, full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes( prepared_data=prepared_data, metric=metric_name ) actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes( assignment_matrix=assignment_matrix, full_cross_disagreement_tensor=full_cross_disagreement_tensor ) assert actual_alpha == pytest.approx(0.811, abs=0.001) class Test_CalcAlphaByPrecomputesFromDictOfDicts(object): def test_e_nominal(self): metric_name = "nominal" with open(os.path.join("tests", "example_E_data.json"), "r") as f: input_table = json.load(f) prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts( input_table=input_table, upper_level="observer", value_constructor=int ) (assignment_matrix, full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes( prepared_data=prepared_data, metric=metric_name ) actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes( assignment_matrix=assignment_matrix, full_cross_disagreement_tensor=full_cross_disagreement_tensor ) assert actual_alpha == pytest.approx(0.743, abs=0.001) def test_e_interval(self): metric_name = "interval" with open(os.path.join("tests", "example_E_data.json"), "r") as f: input_table = json.load(f) prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts( input_table=input_table, upper_level="observer", value_constructor=int ) (assignment_matrix, full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes( prepared_data=prepared_data, metric=metric_name ) actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes( assignment_matrix=assignment_matrix, full_cross_disagreement_tensor=full_cross_disagreement_tensor ) assert actual_alpha == pytest.approx(0.849, abs=0.001) def test_wikipedia_nominal(self): metric_name = "nominal" with open(os.path.join("tests", "example_wikipedia.json"), "r") as f: input_table = json.load(f) prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts( input_table=input_table, upper_level="observer", value_constructor=int ) (assignment_matrix, full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes( prepared_data=prepared_data, metric=metric_name ) actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes( assignment_matrix=assignment_matrix, full_cross_disagreement_tensor=full_cross_disagreement_tensor ) assert actual_alpha == pytest.approx(0.691, abs=0.001) def test_wikipedia_interval(self): metric_name = "interval" with open(os.path.join("tests", "example_wikipedia.json"), "r") as f: input_table = json.load(f) prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts( input_table=input_table, upper_level="observer", value_constructor=int ) (assignment_matrix, full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes( prepared_data=prepared_data, metric=metric_name ) actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes( assignment_matrix=assignment_matrix, full_cross_disagreement_tensor=full_cross_disagreement_tensor ) assert actual_alpha == pytest.approx(0.811, abs=0.001)
41.860051
115
0.628412
2,094
16,451
4.618434
0.065425
0.046324
0.054906
0.060387
0.9301
0.915934
0.909317
0.906111
0.903733
0.903733
0
0.043212
0.2643
16,451
392
116
41.966837
0.755846
0.007781
0
0.740299
0
0
0.042658
0.010542
0
0
0
0
0.065672
1
0.059701
false
0
0.020896
0
0.098507
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
fa77928d2e1c2bfa9e2f7ccbfbd2167d176673e4
4,746
py
Python
modules/MMM-toothbrush/test/BrushSessionTest.py
alexkahn139/MagicMirror
b988d85e833a79ebbdc453b562308056324aa983
[ "MIT" ]
13
2018-12-29T20:01:06.000Z
2021-09-17T09:38:03.000Z
test/BrushSessionTest.py
alexkahn139/OralBlue_python
5430ead8324872691eb5292af15d2e6fbeff9880
[ "Apache-2.0" ]
null
null
null
test/BrushSessionTest.py
alexkahn139/OralBlue_python
5430ead8324872691eb5292af15d2e6fbeff9880
[ "Apache-2.0" ]
8
2019-02-27T06:22:20.000Z
2021-12-14T22:15:19.000Z
import unittest from datetime import datetime, timedelta from OralBlue import OralBAdvertise from OralBlue.BrushMode import BrushMode from OralBlue.BrushSession import BrushSession from OralBlue.BrushState import BrushState class BrushSessionTestCase(unittest.TestCase): def test_anExceptionIsThrownWhenTheDataAreLessThan16Bytes(self): with self.assertRaises(ValueError): BrushSession(b"\x00") def test_anExceptionIsThrownWhenTheDataAreMoreThan16Bytes(self): with self.assertRaises(ValueError): BrushSession(b"\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00") def test_first4byteAreTheStartDate(self): session = BrushSession(b"\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00") self.assertEqual(session.startDate,datetime(year=2000,month=1,day=1)) session = BrushSession(b"\x0A\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00") self.assertEqual(session.startDate, datetime(year=2000, month=1, day=1,second=10)) def test_byte5and6areTheDuration(self): session = BrushSession(b"\x00\x01\x02\x03\x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00") self.assertEqual(session.duration,timedelta(seconds=1)) session = BrushSession(b"\x00\x00\x00\x00\xb4\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00") self.assertEqual(session.duration, timedelta(seconds=180)) session = BrushSession(b"\x00\x00\x00\x00\x00\x20\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00") self.assertEqual(session.duration, timedelta(seconds=0x2000)) def test_byte8IsThePrefMode(self): session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x01\x00\x00\x00\x00\x00\x00\x00\x00") self.assertEqual(session.prefMode,BrushMode(0x01)) session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x04\x00\x00\x00\x00\x00\x00\x00\x00") self.assertEqual(session.prefMode, BrushMode(0x04)) def test_byte9And10IsSecondsUnderPressure(self): session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x00\x00\x00\x00\x00") self.assertEqual(session.timeUnderPressure,timedelta(seconds=0x0908)) session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x01\x00\x00\x00\x00\x00\x00\x00") self.assertEqual(session.timeUnderPressure, timedelta(seconds=1)) def test_byte10IsNumberOfPressure(self): session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x00\x00\x00\x00\x00") self.assertEqual(session.nPressure, 10) session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x00\x00\x00\x00\x00\x00") self.assertEqual(session.nPressure,0) def test_byte11IsBatteryCharge(self): session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x0B\x00\x00\x00\x00") self.assertEqual(session.finalBatteryState, 11) session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x32\x00\x00\x00\x00") self.assertEqual(session.finalBatteryState,50) def test_last4BytesAreTheLastCharge(self): session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x0B\x00\x00\x00\x00") self.assertEqual(session.lastCharge,datetime(year=2000,month=1,day=1)) session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x3B\x0A\x00\x00\x00") self.assertEqual(session.lastCharge,datetime(year=2000, month=1, day=1,second=10)) class BrushSessionV2Or3TestCase(unittest.TestCase): def test_bytes12AsNSectionAndTargetTime(self): session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x0B\x00\x00\x00\x00",protocolVersion=3) self.assertEqual(session.lastCharge,None) self.assertEqual(session.numberOfSector,0) self.assertEqual(session.sessionTargetTime, 0) session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x3B\x78\x80\x00\x00",protocolVersion=3) self.assertEqual(session.lastCharge, None) self.assertEqual(session.numberOfSector, 4) self.assertEqual(session.sessionTargetTime, 120) def test_bytes14AsSessionIdAndUserId(self): session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x0B\x0C\x0D\x00\x00",protocolVersion=3) self.assertEqual(session.lastCharge,None) self.assertEqual(session.sessionId,0) self.assertEqual(session.userId, 0) session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x3B\x78\x80\x03\x20",protocolVersion=3) self.assertEqual(session.lastCharge, None) self.assertEqual(session.sessionId, 3) self.assertEqual(session.userId, 1) if __name__ == '__main__': unittest.main()
53.325843
117
0.722714
657
4,746
5.191781
0.135464
0.228672
0.287599
0.316623
0.734682
0.733216
0.732337
0.722955
0.654647
0.631193
0
0.173723
0.134008
4,746
88
118
53.931818
0.656204
0
0
0.140845
0
0.28169
0.273072
0.270544
0
0
0.004214
0
0.408451
1
0.15493
false
0
0.084507
0
0.267606
0
0
0
0
null
1
1
1
0
1
1
1
0
1
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
null
0
0
0
1
0
0
0
0
0
0
0
0
0
9
fac00226a22a7fdcd7612abc634901a3d3a2425a
162
py
Python
peerplaysapi/wallet.py
farisshajahan/python-peerplays
260e43d39307554f5c35574fe7a6f5451a51697f
[ "MIT" ]
10
2019-03-14T03:09:42.000Z
2021-03-17T10:20:05.000Z
peerplaysapi/wallet.py
farisshajahan/python-peerplays
260e43d39307554f5c35574fe7a6f5451a51697f
[ "MIT" ]
8
2019-04-02T17:07:20.000Z
2020-04-30T08:24:01.000Z
peerplaysapi/wallet.py
farisshajahan/python-peerplays
260e43d39307554f5c35574fe7a6f5451a51697f
[ "MIT" ]
4
2019-04-02T17:00:22.000Z
2021-08-09T11:28:20.000Z
from grapheneapi.grapheneapi import GrapheneAPI class PeerPlaysWalletRPC(GrapheneAPI): """ This class inherits everything from GrapheneAPI """ pass
20.25
55
0.759259
15
162
8.2
0.6
0.243902
0
0
0
0
0
0
0
0
0
0
0.179012
162
7
56
23.142857
0.924812
0.290123
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0.333333
0.333333
0
0.666667
0
1
0
0
null
1
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
1
1
0
1
0
0
7
fad2774df466156d62a3a38ab65ff7e2b6897589
146
py
Python
angrcli/full/__init__.py
cokesme/angr-cli-mirror
37da5bf731549dab2523215a9212f372e3302168
[ "MIT" ]
38
2018-10-18T17:03:52.000Z
2022-02-06T13:19:19.000Z
angrcli/full/__init__.py
cokesme/angr-cli-mirror
37da5bf731549dab2523215a9212f372e3302168
[ "MIT" ]
3
2019-06-29T10:16:20.000Z
2021-06-10T15:29:27.000Z
angrcli/full/__init__.py
cokesme/angr-cli-mirror
37da5bf731549dab2523215a9212f372e3302168
[ "MIT" ]
4
2018-10-08T12:54:50.000Z
2020-03-27T09:29:27.000Z
from angrcli.plugins.watches import Watches from angrcli.plugins.ContextView import ContextView from angrcli.plugins.explore import ExplorePlugin
36.5
51
0.876712
18
146
7.111111
0.444444
0.257813
0.421875
0
0
0
0
0
0
0
0
0
0.082192
146
3
52
48.666667
0.955224
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
fada392eade7c3cd7bc3b09d1b80c50aa2b5674b
44,679
py
Python
tests/integration_tests/test_resources/new_test_ads.py
JobtechSwe/elastic-importers
ae370984f79295a784350f98e695977a1f73647a
[ "Apache-2.0" ]
2
2020-03-02T18:29:08.000Z
2021-06-09T00:48:24.000Z
tests/integration_tests/test_resources/new_test_ads.py
JobtechSwe/elastic-importers
ae370984f79295a784350f98e695977a1f73647a
[ "Apache-2.0" ]
12
2019-01-31T09:54:23.000Z
2021-10-04T11:25:44.000Z
tests/integration_tests/test_resources/new_test_ads.py
JobtechSwe/elastic-importers
ae370984f79295a784350f98e695977a1f73647a
[ "Apache-2.0" ]
1
2018-11-09T14:44:55.000Z
2018-11-09T14:44:55.000Z
test_ads = [ { "annonsId": 23448227, "annonsrubrik": "Driven fältsäljare Östergötland/Västergötland", "annonstext": "Köksbörsen är inne i ett expansivt skede. Vi har egen import ifrån Europa och Asien samt USA. Vi har egna agenturprodukter. Vi levererar och projekterar kompletta restaurangkök över hela Sverige. Vi har en säljare i Göteborg, en säljare i Umeå/Skellefteå samt ett antal säljare i Stockholm. Vi expanderar nu säljteamet i Stockholm med ytterligare en fältsäljare för Stockholm/Mälardalen, då vi har sett ett utökat behov av ytterligare en säljare, vi ett härligt säljteam!\nVi ligger centralt i Stockholm vid Gärdet/Östermalm. VI har egen utställning av nya/begagnade restaurangmaskiner, samt även en utställning för möbler/glas/porslin/bestick. För att lyckas med denna tjänst, så krävs flera års erfarenhet inom försäljning till dagligvaruhandeln/restaurangbranschen, det krävs att du har ett brett kontaktnät inom branschen och många kontakter att börja jobba med direkt. Det är ett stort plus om du tidigare har jobbat som säljare av restaurangmaskiner/inredning till restauranger/dagligvaruhandeln. I den här rollen kommer fokus att ligga på nykundsbearbetning, men du kommer även att utveckla befintliga kunder. Du ansvarar för din egen budget, kundstock och du bokar dina egna möten i ditt arbete.\nDet är meriterande om du har goda försäljningsresultat sedan tidigare och det är ett krav att du har körkort. För att lyckas i rollen bör du tycka om att arbeta med uppsökande försäljning och vara duktig på att skapa ett brett kontaktnät inom branschen. Du ska självständigt kunna planera ditt arbete och vara en driven person som är hungrig på nya utmaningar.\nDu är en självgående person som kan ta egna initiativ till nya affärer. Som person är du driven, stresstålig och ansvarsfull. Du har skinn på näsan och kan hantera många saker samtidigt. Det är viktigt att du är målinriktad och resultatfokuserad i ditt arbete och vill få snabba avslut i dina affärer. Vi ser att du är en social person som tycker om att ha mycket kundkontakt samt har ett sinne för att göra affärer. Vi ser det som meriterande om du har CAD vana. En fördel är om du är utbildad kock, eller har jobbat på restaurang. Vi söker en driven fältsäljare som kommer att bearbeta nya samt befintliga kunder i Östergötland/Västergötland. Du ansvarar för din egen budget, kundstock och du bokar dina egna möten. I rollen som fältsäljare planerar du dina egna dagar och det är därför viktigt att du är ansvarsfull och målinriktad i ditt arbete.\nFör att lyckas med denna tjänst, så krävs flera års erfarenhet inom försäljning till dagligvaruhandeln/restaurangbranschen, det krävs att du har ett brett kontaktnät inom branschen och många kontakter att börja jobba med direkt. Du kommer även jobba med ringlistor för att bearbeta kedjekunder samt byggprojekt. Dom kunder som du ska bearbeta är restaurangkedjor, hotel, byggfirmor, kökskonsulter/arkitekter, livsmedelsbutiker, bensinstationer mm.", "annonstextFormaterad": "Köksbörsen är inne i ett expansivt skede. Vi har egen import ifrån Europa och Asien samt USA. Vi har egna agenturprodukter. Vi levererar och projekterar kompletta restaurangkök över hela Sverige. Vi har en säljare i Göteborg, en säljare i Umeå/Skellefteå samt ett antal säljare i Stockholm. Vi expanderar nu säljteamet i Stockholm med ytterligare en fältsäljare för Stockholm/Mälardalen, då vi har sett ett utökat behov av ytterligare en säljare, vi ett härligt säljteam!<br>Vi ligger centralt i Stockholm vid Gärdet/Östermalm. VI har egen utställning av nya/begagnade restaurangmaskiner, samt även en utställning för möbler/glas/porslin/bestick. För att lyckas med denna tjänst, så krävs flera års erfarenhet inom försäljning till dagligvaruhandeln/restaurangbranschen, det krävs att du har ett brett kontaktnät inom branschen och många kontakter att börja jobba med direkt. Det är ett stort plus om du tidigare har jobbat som säljare av restaurangmaskiner/inredning till restauranger/dagligvaruhandeln. I den här rollen kommer fokus att ligga på nykundsbearbetning, men du kommer även att utveckla befintliga kunder. Du ansvarar för din egen budget, kundstock och du bokar dina egna möten i ditt arbete.<br>Det är meriterande om du har goda försäljningsresultat sedan tidigare och det är ett krav att du har körkort. För att lyckas i rollen bör du tycka om att arbeta med uppsökande försäljning och vara duktig på att skapa ett brett kontaktnät inom branschen. Du ska självständigt kunna planera ditt arbete och vara en driven person som är hungrig på nya utmaningar.<br>Du är en självgående person som kan ta egna initiativ till nya affärer. Som person är du driven, stresstålig och ansvarsfull. Du har skinn på näsan och kan hantera många saker samtidigt. Det är viktigt att du är målinriktad och resultatfokuserad i ditt arbete och vill få snabba avslut i dina affärer. Vi ser att du är en social person som tycker om att ha mycket kundkontakt samt har ett sinne för att göra affärer. Vi ser det som meriterande om du har CAD vana. En fördel är om du är utbildad kock, eller har jobbat på restaurang. Vi söker en driven fältsäljare som kommer att bearbeta nya samt befintliga kunder i Östergötland/Västergötland. Du ansvarar för din egen budget, kundstock och du bokar dina egna möten. I rollen som fältsäljare planerar du dina egna dagar och det är därför viktigt att du är ansvarsfull och målinriktad i ditt arbete.<br>För att lyckas med denna tjänst, så krävs flera års erfarenhet inom försäljning till dagligvaruhandeln/restaurangbranschen, det krävs att du har ett brett kontaktnät inom branschen och många kontakter att börja jobba med direkt. Du kommer även jobba med ringlistor för att bearbeta kedjekunder samt byggprojekt. Dom kunder som du ska bearbeta är restaurangkedjor, hotel, byggfirmor, kökskonsulter/arkitekter, livsmedelsbutiker, bensinstationer mm.", "ansokningssattEpost": " EMAIL ", "ansokningssattViaAF": False, "anstallningTyp": { "namn": "Vanlig anställning", "varde": "PFZr_Syz_cUq" }, "antalPlatser": 1, "arbetsgivareId": "20513633", "arbetsgivareNamn": "Köksbörsen AB", "arbetsplatsId": "83191774", "arbetsplatsNamn": "KÖKSBÖRSEN AB", "arbetsplatsadress": { "gatuadress": "Linköping/Jönköping", "kommun": { "namn": "Linköping", "varde": "bm2x_1mr_Qhx" }, "koordinatPrecision": "POSTNUMMER", "lan": { "namn": "Östergötlands län", "varde": "oLT3_Q9p_3nn" }, "land": { "namn": "Sverige", "varde": "i46j_HmG_v64" }, "latitud": "58.4044001929687", "longitud": "15.6003742645864", "postnr": "58002", "postort": "LINKÖPING" }, "arbetstidTyp": { "namn": "Heltid", "varde": "6YE1_gAC_R2G" }, "avpublicerad": False, "ejKravSvenskaEngelska": False, "informationAnsokningssatt": None, "ingenErfarenhetKravs": False, "inkluderande": False, "kallaTyp": "VIA_ANNONSERA", "kontaktpersoner": [ { "befattning": None, "beskrivning": "blabla", "efternamn": "Testsson", "epost": "test@jobtechdev.se", "fackligRepresentant": False, "fornamn": "Testy", "telefonnummer": "+01011122233" } ], "lonTyp": { "namn": "Fast och rörlig lön", "varde": "asrX_9Df_ukn" }, "organisationsnummer": "5567425631", "publiceringsdatum": "2021-03-19 12:41:52", "referens": "Utesäljare", "sistaAnsokningsdatum": "2021-04-18 23:59:59", "sistaPubliceringsdatum": "2022-06-14 07:38:46", "telefonnummer": None, "tillgangTillEgenBil": False, "uppdateradTid": 1616154112141, "utbildningsinriktning": { "namn": "Företagsekonomi, handel och administration", "varde": "CU11_PU7_Yj1", "vikt": 5 }, "utbildningsniva": { "namn": "Gymnasial utbildning", "varde": "DeBt_ahh_bkx", "vikt": 5 }, "varaktighetTyp": { "namn": "Tillsvidare", "varde": "a7uU_j21_mkL" }, "version": "1.0", "yrkeserfarenheter": [ { "erfarenhetsniva": { "namn": "1-2 års erfarenhet", "varde": "LLnd_5GJ_4ju" }, "namn": "Utesäljare", "varde": "p17k_znk_osi", "vikt": 10 } ], "yrkesroll": { "namn": "Utesäljare", "varde": "p17k_znk_osi" } }, {'annonsId': 23699999, 'annonsrubrik': 'Greenkeeper -- Banpersonal 2021 Ullna Golf', 'annonstext': 'Ullna Golf Club är en klassisk golfanläggning som funnits sedan 1981, vi har som mål att presentera en förstklassig golfbana och erbjuda en service med hög nivå.\n\nDu som söker dig till oss är positiv, flexibel, ordningsam och intresserad av att jobba i grupp. Du kommer jobba med skötsel av golfbanan. Det är tidiga morgnar, runt 06.00, i gengäld slutar vi tidigt på eftermiddagen, ca 15.00, du kommer även jobba några helger.\n\nDina arbetsuppgifter kommer till största del bestå av att klippa tees (utslagsplatser), greener (mest finklippta ytan), klippa ruffar samt kratta bunkrar.\n\nVi söker ett flertal banarbetare med varierande start av tjänsterna. De första tjänsterna börjar i FIRSTNAME/LASTNAME och löper i ca 4-6 månader beroende på tjänst. Det är ett krav att du kan ta dig hit själv, vi har mycket dåliga kollektiva kommunikationer. Du måste även kunna göra dig förstådd och förstå engelska. Det är meriterande om du spelar golf och/eller har jobbat på en golfbana, har B-kort eller traktorbehörighet. Det är bra om du har jobbat med röj såg, kört traktor.\n\nVi ser fram emot att läsa din ansökan och i slutändan jobba tillsammans.\n\nRekryteringen sker löpande.\n\nFör frågor och ansökningar mailas till \n\n Banchef FIRSTNAME LASTNAME\n\n EMAIL \n\nSkriv "Banpersonal2021" i ämnesraden\n\nDu kan besöka oss på Ullnagolf.se', 'annonstextFormaterad': '<p>Ullna Golf Club är en klassisk golfanläggning som funnits sedan 1981, vi har som mål att presentera en förstklassig golfbana och erbjuda en service med hög nivå.</p><p>Du som söker dig till oss är positiv, flexibel, ordningsam och intresserad av att jobba i grupp. Du kommer jobba med skötsel av golfbanan. Det är tidiga morgnar, runt 06.00, i gengäld slutar vi tidigt på eftermiddagen, ca 15.00, du kommer även jobba några helger.</p><p>Dina arbetsuppgifter kommer till största del bestå av att klippa tees (utslagsplatser), greener (mest finklippta ytan), klippa ruffar samt kratta bunkrar.</p><p>Vi söker ett flertal banarbetare med varierande start av tjänsterna. De första tjänsterna börjar i FIRSTNAME/LASTNAME och löper i ca 4-6 månader beroende på tjänst. Det är ett krav att du kan ta dig hit själv, vi har mycket dåliga kollektiva kommunikationer. Du måste även kunna göra dig förstådd och förstå engelska. Det är meriterande om du spelar golf och/eller har jobbat på en golfbana, har B-kort eller traktorbehörighet. Det är bra om du har jobbat med röj såg, kört traktor.</p><p>Vi ser fram emot att läsa din ansökan och i slutändan jobba tillsammans.</p><p>Rekryteringen sker löpande.</p><p><br></p><p>För frågor och ansökningar mailas till </p><p> Banchef FIRSTNAME LASTNAME</p><p> EMAIL </p><p>Skriv "Banpersonal2021" i ämnesraden</p><p>Du kan besöka oss på Ullnagolf.se</p><p><br></p><p><br></p>', 'ansokningssattEpost': ' EMAIL ', 'ansokningssattViaAF': False, 'anstallningTyp': {'namn': 'Sommarjobb / feriejobb', 'varde': 'Jh8f_q9J_pbJ'}, 'antalPlatser': 4, 'arbetsgivareId': '21057934', 'arbetsgivareNamn': 'Ullna Golf AB', 'arbetsplatsId': '85992110', 'arbetsplatsNamn': 'Ullna Golf AB', 'arbetsplatsadress': {'gatuadress': 'Roslagsvägen 36', 'kommun': {'namn': 'Österåker', 'varde': '8gKt_ZsV_PGj'}, 'koordinatPrecision': 'GATUADRESS', 'lan': {'namn': 'Stockholms län', 'varde': 'CifL_Rzy_Mku'}, 'land': {'namn': 'Sverige', 'varde': 'i46j_HmG_v64'}, 'latitud': '59.49548570116787', 'longitud': '18.152517169298417', 'postnr': '18494', 'postort': 'Åkersberga'}, 'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'ejKravSvenskaEngelska': False, 'informationAnsokningssatt': None, 'ingenErfarenhetKravs': False, 'inkluderande': False, 'kallaTyp': 'VIA_ANNONSERA', 'kontaktpersoner': [ {'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se', 'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}], 'korkort': [{'namn': 'B', 'varde': 'VTK8_WRx_GcM'}], 'lonTyp': {'namn': 'Fast månads- vecko- eller timlön', 'varde': 'oG8G_9cW_nRf'}, 'organisationsnummer': '5560428095', 'publiceringsdatum': '2021-03-18 00:00:00', 'referens': 'Banpersonal 2021', 'sistaAnsokningsdatum': '2021-04-16 23:59:59', 'sistaPubliceringsdatum': '2022-06-14 14:33:12', 'telefonnummer': None, 'tillgangTillEgenBil': True, 'uppdateradTid': 1615971812090, 'varaktighetTyp': {'namn': '3 - 6 månader', 'varde': 'Xj7x_7yZ_jEn'}, 'version': '1.0', 'webbadress': 'ullnagolf.se', 'yrkesroll': {'namn': 'Golfbanearbetare', 'varde': 'tArV_EVU_cFQ'}}, {'annonsId': 24277938, 'annonsrubrik': 'Sjuksköterska till vårdplatsenheten', 'annonstext': 'Kullbergska sjukhuset Katrineholm \r\n\r\nVälkommen att skapa Sveriges friskaste län tillsammans med oss!\n\nSjuksköterska till vårdplatsenheten, se hit!\n\nOm oss\nVårdplatsenheten har två vårdavdelningar varav den ena med inriktning hjärt -och medicin och den andra stroke- och medicin. Vårdplatsenheten består av ett härligt gäng medarbetare som nu behöver utöka sjuksköterskegruppen. \nAvdelningarna vårdar patienter med olika sjukdomstillstånd, vi söker dig som är intresserad av hjärtsjukvård och strokevård/ortopedrehab.\nDu ska vara nyfiken, flexibel och kunna anpassa dig till ett varierande arbete och tempo.\n\nVi jobbar i team bestående av sjuksköterskor, undersköterskor, utskrivningssamordnare och medicinska sekreterare. Läkare och paramedicinsk kompetens finns på avdelningen under vissa tider som stöd i arbetet. \n\nTillsammans arbetar vi målinriktat för att skapa en trygg och säker resa genom vården för våra kunder.\n\nDin kompetens\nLegitimerad sjuksköterska. Erfarenhet av slutenvård är meriterande men inget krav. Vi söker en flexibel person som uppskattar en varierad arbetsdag med möjlighet att lära sig mycket i arbetet. Du ska ha lätt att samarbeta med olika professioner, och vara öppen för förändringar. Vi lägger stor vikt på att vara lyhörd och göra insatser där behovet är som störst. Personliga egenskaper lägger vi stor vikt vid.\n\nAnställningsform\nTillsvidareanställning / Vikariat på heltid eller enligt överenskommelse. Tillträde enligt överenskommelse.\nArbetstid dag/kväll/helg eller natt. Planeringsschema används på avdelningen med möjlighet att kunna påverka sina pass till viss del.\n\nInformation om tjänsten lämnas av\nVårdenhetschef FIRSTNAME LASTNAME (stroke- och medicin), TELEPHONENO .\nVårdenhetschef Ann-FIRSTNAME LASTNAME (hjärtmedicin), TELEPHONENO .\nFacklig företrädare Arleide Pereira FIRSTNAME LASTNAME, TELEPHONENO och FIRSTNAME LASTNAME, TELEPHONENO .\nÖvriga fackliga företrädare nås via kontaktcenter, TELEPHONENO .\n\nKom och jobba hos oss på vårdplatsenheten!\n\nVälkommen med din ansökan, inklusive CV, senast 2021-03-29.\nIntervjuer kan komma att ske löpande.\n\nSe våra förmåner (http://regionsormland.se/jobb-och-utbildning/vara-formaner/)\n\nFölj oss gärna på Facebook (http://facebook.com/regionsormlandJobbahososs/)\n\r\n\nVi har i vissa fall skyldighet att kontrollera om en person förekommer i misstanke- och eller belastningsregistret. Det kan ske på två sätt, endera begär regionen ut uppgiften själv eller också uppmanas du att begära ut utdrag för att kunna visa upp innan anställning. Vi begär i undantagsfall att du visar upp registerutdrag även vid tillsättning av andra tjänster än de som avses ovan. Blir du aktuell för anställning kommer du att informeras om vad som gäller för den tjänst du sökt.\n\nVi gör aktiva val vid exponering och rekryteringsstöd och undanber oss därför direktkontakt av bemannings- och rekryteringsföretag.', 'annonstextFormaterad': 'Kullbergska sjukhuset Katrineholm \r\n\r\nVälkommen att skapa Sveriges friskaste län tillsammans med oss!\n\nSjuksköterska till vårdplatsenheten, se hit!\n\nOm oss\nVårdplatsenheten har två vårdavdelningar varav den ena med inriktning hjärt -och medicin och den andra stroke- och medicin. Vårdplatsenheten består av ett härligt gäng medarbetare som nu behöver utöka sjuksköterskegruppen. \nAvdelningarna vårdar patienter med olika sjukdomstillstånd, vi söker dig som är intresserad av hjärtsjukvård och strokevård/ortopedrehab.\nDu ska vara nyfiken, flexibel och kunna anpassa dig till ett varierande arbete och tempo.\n\nVi jobbar i team bestående av sjuksköterskor, undersköterskor, utskrivningssamordnare och medicinska sekreterare. Läkare och paramedicinsk kompetens finns på avdelningen under vissa tider som stöd i arbetet. \n\nTillsammans arbetar vi målinriktat för att skapa en trygg och säker resa genom vården för våra kunder.\n\nDin kompetens\nLegitimerad sjuksköterska. Erfarenhet av slutenvård är meriterande men inget krav. Vi söker en flexibel person som uppskattar en varierad arbetsdag med möjlighet att lära sig mycket i arbetet. Du ska ha lätt att samarbeta med olika professioner, och vara öppen för förändringar. Vi lägger stor vikt på att vara lyhörd och göra insatser där behovet är som störst. Personliga egenskaper lägger vi stor vikt vid.\n\nAnställningsform\nTillsvidareanställning / Vikariat på heltid eller enligt överenskommelse. Tillträde enligt överenskommelse.\nArbetstid dag/kväll/helg eller natt. Planeringsschema används på avdelningen med möjlighet att kunna påverka sina pass till viss del.\n\nInformation om tjänsten lämnas av\nVårdenhetschef FIRSTNAME LASTNAME (stroke- och medicin), TELEPHONENO .\nVårdenhetschef Ann-FIRSTNAME LASTNAME (hjärtmedicin), TELEPHONENO .\nFacklig företrädare Arleide Pereira FIRSTNAME LASTNAME, TELEPHONENO och FIRSTNAME LASTNAME, TELEPHONENO .\nÖvriga fackliga företrädare nås via kontaktcenter, TELEPHONENO .\n\n\nKom och jobba hos oss på vårdplatsenheten!\n\n\nVälkommen med din ansökan, inklusive CV, senast 2021-03-29.\nIntervjuer kan komma att ske löpande.\n\nSe våra förmåner (http://regionsormland.se/jobb-och-utbildning/vara-formaner/)\n\nFölj oss gärna på Facebook (http://facebook.com/regionsormlandJobbahososs/)\n\r\n\nVi har i vissa fall skyldighet att kontrollera om en person förekommer i misstanke- och eller belastningsregistret. Det kan ske på två sätt, endera begär regionen ut uppgiften själv eller också uppmanas du att begära ut utdrag för att kunna visa upp innan anställning. Vi begär i undantagsfall att du visar upp registerutdrag även vid tillsättning av andra tjänster än de som avses ovan. Blir du aktuell för anställning kommer du att informeras om vad som gäller för den tjänst du sökt.\n\nVi gör aktiva val vid exponering och rekryteringsstöd och undanber oss därför direktkontakt av bemannings- och rekryteringsföretag.', 'ansokningsadress': {'gatuadress': 'Repslagaregatan 19', 'land': None, 'mottagare': None, 'postnr': None, 'postort': None}, 'ansokningssattEpost': None, 'ansokningssattViaAF': False, 'ansokningssattWebbadress': 'https://sormland.powerinit.com/Modules/Recruitments/Public/?JobPositionId=18351&RefNo=RLSV-20-564&Source=[JobPositionSource]', 'anstallningTyp': {'namn': 'Vanlig anställning', 'varde': 'PFZr_Syz_cUq'}, 'antalPlatser': 1, 'arbetsgivareId': '10820079', 'arbetsgivareNamn': 'REGION SÖRMLAND', 'arbetsplatsId': '0', 'arbetsplatsNamn': 'Region Sörmland', 'arbetsplatsadress': {'gatuadress': None, 'kommun': {'namn': 'Katrineholm', 'varde': 'snx9_qVD_Dr1'}, 'koordinatPrecision': 'KOMMUN', 'lan': {'namn': 'Södermanlands län', 'varde': 's93u_BEb_sx2'}, 'land': {'namn': 'Sverige', 'varde': 'i46j_HmG_v64'}, 'latitud': '58.995552', 'longitud': '16.205475', 'postnr': None, 'postort': None}, 'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'besoksadress': {'gatuadress': 'Repslagaregatan 19', 'land': None, 'postnr': None, 'postort': None}, 'ejKravSvenskaEngelska': False, 'externtAnnonsId': '46-232100-0032-18351', 'informationAnsokningssatt': None, 'ingenErfarenhetKravs': False, 'inkluderande': False, 'kallaTyp': 'VIA_PLATSBANKEN_DXA', 'kontaktpersoner': [ {'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se', 'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}], 'lonTyp': {'namn': 'Fast månads- vecko- eller timlön', 'varde': 'oG8G_9cW_nRf'}, 'lonebeskrivning': '-', 'organisationsnummer': '2321000032', 'postadress': {'gatuadress': 'Repslagaregatan 19', 'land': 'SE', 'postnr': '61188', 'postort': 'Nyköping'}, 'publiceringsdatum': '2020-09-30 14:35:20', 'referens': 'RLSV-20-564', 'sistaAnsokningsdatum': '2021-03-29 23:59:59', 'sistaPubliceringsdatum': '2022-06-14 14:33:59', 'telefonnummer': None, 'tillgangTillEgenBil': False, 'uppdateradTid': 1613993432534, 'varaktighetTyp': {'namn': 'Tillsvidare', 'varde': 'a7uU_j21_mkL'}, 'version': '4.0', 'villkorsbeskrivning': 'Tillsvidare, 100, Tillträde enligt överenskommelse\r\n\n-', 'yrkeserfarenheter': [ {'erfarenhetsniva': {'namn': 'Mindre än 1 års erfarenhet', 'varde': 'yrAe_Fzi_E6u'}, 'namn': 'Sjuksköterska, grundutbildad', 'varde': 'bXNH_MNX_dUR', 'vikt': 10}], 'yrkesroll': {'namn': 'Sjuksköterska, grundutbildad', 'varde': 'bXNH_MNX_dUR'}}] ads_not_sweden = [ {'annonsId': 24458381, 'annonsrubrik': 'Servicetekniker i Oslo sökes', 'annonstext': 'Er du en handy og serviceinnstilt person som ønsker en spennende stilling i en fremoverlent og voksende bedrift?\n\nSom servicetekniker vil du ha ansvar for service og hjemmebesøk for Part Construction AB. Du vil behandle innkommende forespørsler, utføre service oppdrag på byggeprosjekter og reklamasjonsarbeid. Arbeidet foregår hovedsakelig på Østlandet, men vi har også leveranser til andre deler av landet. Noe reising vil derfor måtte påberegnes.\n\nOm deg:\n\n\tEn allsidig praktisk fagarbeider. Du er kanskje flislegger, rørlegger, elektriker eller har en annen relevant utdanning\n\tSnakker og skriver godt Norsk\n\tFørerkort klasse B/BE\n\tGod evne til kundebehandling\n\tEr selvstendig og strukturert\n\tGrunnleggende data kunnskaper.\n\tKjennskap til NS 3420, NS 8409 og Bustadoppføringslova.\n\nPart Construction AB kan tilby:\n\n\tEt dynamisk og spennende arbeidsmiljø\n\tUtfordrende og variert arbeidsoppgaver\n\tKonkurransedyktig betingelser\n\tServicebil, mobil telefon\n\tGode forsikringer og pensjons ordninger\n\nOm arbeidsgiveren \n\nPart Construction AB, en av Europas ledende produsenter av prefabrikkerte baderoms moduler. Part Construction AB tilbyr entreprenørfirmaer og eiendomsselskaper prefabrikkerte baderom til blant annet, hoteller, sykehjem og boliger. Bademodulene leveres komplett med fliser, gulv, møbler, inventar, helt etter kundens ønsker. rørleggerarbeid og elektriske installasjoner er også ferdig montert fra fabrikk, Part er et familieeid selskap fra Kalix, med ca. 300 ansatte og med en omsetning på ca. 600 millioner. Hovedkvarter og produksjon er i Kalix / Överkalix, Norrbotten. Part Construction AB er en del av konserngruppen PartGroup sammen med søsterselskapene PreBad AB, Altor Industri, Isolamin AB, Space Interior AB, og PCS Modulsystem.\n\nSektor\n\nPrivat\n\nSted\n\nVestbygata 55 2003 Lillestrøm\n\nBransje\n\nBygg og anlegg,\n\nStillingsfunksjon\n\nHåndverker / Andre montører,\n\nHåndverker,\n\nHåndverker / Flislegger, rørlegger, elektriker\n\n \n\nSøknad\n\nTiltredelse\n\nEtter avtale\n\nVarighet\n\nFast, heltid (6 måneders prøvetid)\n\nBosted\n\nOslo området\n\nSøknadsfrist\n\nSnarest, og innen 10.03.2021\n\nSpørsmål om stillingen rettes til FIRSTNAME LASTNAME Yttervik, Servicekoordinator Norge. Tel: TELEPHONENO , e-post: EMAIL \n\nSend søknad merket "Servicetekniker Norge" på e-post til: EMAIL \n\nDenne rekrutteringen skjer helt i Part Constructions regi, og vi ønsker kun direkte kontakt med personlig søkere. \n\nVi gleder oss til å høre fra deg!', 'annonstextFormaterad': '<p><strong>Er du en handy og serviceinnstilt person som ønsker en spennende stilling i en fremoverlent og voksende bedrift?</strong></p><p>Som servicetekniker vil du ha ansvar for service og hjemmebesøk for Part Construction AB. Du vil behandle innkommende forespørsler, utføre service oppdrag på byggeprosjekter og reklamasjonsarbeid. Arbeidet foregår hovedsakelig på Østlandet, men vi har også leveranser til andre deler av landet. Noe reising vil derfor måtte påberegnes.</p><p><br></p><p><strong>Om deg:</strong></p><ul><li>En allsidig praktisk fagarbeider. Du er kanskje flislegger, rørlegger, elektriker eller har en annen relevant utdanning</li><li>Snakker og skriver godt Norsk</li><li>Førerkort klasse B/BE</li><li>God evne til kundebehandling</li><li>Er selvstendig og strukturert</li><li>Grunnleggende data kunnskaper.</li><li>Kjennskap til NS 3420, NS 8409 og Bustadoppføringslova.</li></ul><p><br></p><p><strong>Part Construction AB kan tilby:</strong></p><ul><li>Et dynamisk og spennende arbeidsmiljø</li><li>Utfordrende og variert arbeidsoppgaver</li><li>Konkurransedyktig betingelser</li><li>Servicebil, mobil telefon</li><li>Gode forsikringer og pensjons ordninger</li></ul><p><br></p><p><strong>Om arbeidsgiveren </strong></p><p>Part Construction AB, en av Europas ledende produsenter av prefabrikkerte baderoms moduler. Part Construction AB tilbyr entreprenørfirmaer og eiendomsselskaper prefabrikkerte baderom til blant annet, hoteller, sykehjem og boliger. Bademodulene leveres komplett med fliser, gulv, møbler, inventar, helt etter kundens ønsker. rørleggerarbeid og elektriske installasjoner er også ferdig montert fra fabrikk, Part er et familieeid selskap fra Kalix, med ca. 300 ansatte og med en omsetning på ca. 600 millioner. Hovedkvarter og produksjon er i Kalix / Överkalix, Norrbotten. Part Construction AB er en del av konserngruppen PartGroup sammen med søsterselskapene PreBad AB, Altor Industri, Isolamin AB, Space Interior AB, og PCS Modulsystem.</p><p><br></p><p><strong>Sektor</strong></p><p>Privat</p><p><strong>Sted</strong></p><p>Vestbygata 55 2003 Lillestrøm</p><p><strong>Bransje</strong></p><p>Bygg og anlegg,</p><p><strong>Stillingsfunksjon</strong></p><p>Håndverker / Andre montører,</p><p>Håndverker,</p><p>Håndverker / Flislegger, rørlegger, elektriker</p><p> </p><p><strong><u>Søknad</u></strong></p><p><br></p><p><strong>Tiltredelse</strong></p><p>Etter avtale</p><p><strong>Varighet</strong></p><p>Fast, heltid (6 måneders prøvetid)</p><p><strong>Bosted</strong></p><p>Oslo området</p><p><strong>Søknadsfrist</strong></p><p>Snarest, og innen 10.03.2021</p><p><strong>Spørsmål </strong>om stillingen rettes til FIRSTNAME LASTNAME Yttervik, Servicekoordinator Norge. Tel: TELEPHONENO , e-post: EMAIL </p><p>Send søknad merket "Servicetekniker Norge" på e-post til: EMAIL </p><p>Denne rekrutteringen skjer helt i Part Constructions regi, og vi ønsker kun direkte kontakt med personlig søkere. </p><p>Vi gleder oss til å høre fra deg!</p><p id="oppenforalla">Öppen för alla\nVi fokuserar på din kompetens, inte dina övriga förutsättningar. Vi är öppna för att anpassa rollen eller arbetsplatsen efter dina behov.</p>', 'ansokningssattEpost': ' EMAIL ', 'ansokningssattViaAF': False, 'anstallningTyp': {'namn': 'Vanlig anställning', 'varde': 'PFZr_Syz_cUq'}, 'antalPlatser': 1, 'arbetsgivareId': '20610987', 'arbetsgivareNamn': 'Partbyggen i Kalix AB', 'arbetsplatsId': '86440232', 'arbetsplatsNamn': 'Part Construction', 'arbetsplatsadress': {'gatuadress': None, 'kommun': None, 'koordinatPrecision': None, 'lan': None, 'land': {'namn': 'Norge', 'varde': 'QJgN_Zge_BzJ'}, 'latitud': None, 'longitud': None, 'postnr': None, 'postort': None}, 'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'ejKravSvenskaEngelska': False, 'informationAnsokningssatt': None, 'ingenErfarenhetKravs': False, 'inkluderande': True, 'kallaTyp': 'VIA_ANNONSERA', 'kontaktpersoner': [ {'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se', 'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}], 'korkort': [{'namn': 'B', 'varde': 'VTK8_WRx_GcM'}], 'lonTyp': {'namn': 'Fast månads- vecko- eller timlön', 'varde': 'oG8G_9cW_nRf'}, 'organisationsnummer': '5564003357', 'publiceringsdatum': '2021-01-04 16:18:13', 'referens': 'Servicetekniker Norge', 'sistaAnsokningsdatum': '2021-03-31 23:59:59', 'sistaPubliceringsdatum': '2022-06-14 14:52:55', 'sprak': [{'namn': 'Norska', 'varde': 'pnjj_2JX_Fub', 'vikt': 10}], 'telefonnummer': None, 'tillgangTillEgenBil': False, 'uppdateradTid': 1609773493583, 'varaktighetTyp': {'namn': 'Tillsvidare', 'varde': 'a7uU_j21_mkL'}, 'version': '1.0', 'webbadress': 'partab.nu', 'yrkeserfarenheter': [ {'erfarenhetsniva': {'namn': '1-2 års erfarenhet', 'varde': 'LLnd_5GJ_4ju'}, 'namn': 'VVS-montör', 'varde': 'ZZTi_v6g_4cZ', 'vikt': 5}], 'yrkesroll': {'namn': 'VVS-montör', 'varde': 'ZZTi_v6g_4cZ'}}, {'annonsId': 24458381, 'annonsrubrik': 'Servicetekniker i Oslo sökes', 'annonstext': 'Er du en handy og serviceinnstilt person som ønsker en spennende stilling i en fremoverlent og voksende bedrift?\n\nSom servicetekniker vil du ha ansvar for service og hjemmebesøk for Part Construction AB. Du vil behandle innkommende forespørsler, utføre service oppdrag på byggeprosjekter og reklamasjonsarbeid. Arbeidet foregår hovedsakelig på Østlandet, men vi har også leveranser til andre deler av landet. Noe reising vil derfor måtte påberegnes.\n\nOm deg:\n\n\tEn allsidig praktisk fagarbeider. Du er kanskje flislegger, rørlegger, elektriker eller har en annen relevant utdanning\n\tSnakker og skriver godt Norsk\n\tFørerkort klasse B/BE\n\tGod evne til kundebehandling\n\tEr selvstendig og strukturert\n\tGrunnleggende data kunnskaper.\n\tKjennskap til NS 3420, NS 8409 og Bustadoppføringslova.\n\nPart Construction AB kan tilby:\n\n\tEt dynamisk og spennende arbeidsmiljø\n\tUtfordrende og variert arbeidsoppgaver\n\tKonkurransedyktig betingelser\n\tServicebil, mobil telefon\n\tGode forsikringer og pensjons ordninger\n\nOm arbeidsgiveren \n\nPart Construction AB, en av Europas ledende produsenter av prefabrikkerte baderoms moduler. Part Construction AB tilbyr entreprenørfirmaer og eiendomsselskaper prefabrikkerte baderom til blant annet, hoteller, sykehjem og boliger. Bademodulene leveres komplett med fliser, gulv, møbler, inventar, helt etter kundens ønsker. rørleggerarbeid og elektriske installasjoner er også ferdig montert fra fabrikk, Part er et familieeid selskap fra Kalix, med ca. 300 ansatte og med en omsetning på ca. 600 millioner. Hovedkvarter og produksjon er i Kalix / Överkalix, Norrbotten. Part Construction AB er en del av konserngruppen PartGroup sammen med søsterselskapene PreBad AB, Altor Industri, Isolamin AB, Space Interior AB, og PCS Modulsystem.\n\nSektor\n\nPrivat\n\nSted\n\nVestbygata 55 2003 Lillestrøm\n\nBransje\n\nBygg og anlegg,\n\nStillingsfunksjon\n\nHåndverker / Andre montører,\n\nHåndverker,\n\nHåndverker / Flislegger, rørlegger, elektriker\n\n \n\nSøknad\n\nTiltredelse\n\nEtter avtale\n\nVarighet\n\nFast, heltid (6 måneders prøvetid)\n\nBosted\n\nOslo området\n\nSøknadsfrist\n\nSnarest, og innen 10.03.2021\n\nSpørsmål om stillingen rettes til FIRSTNAME LASTNAME Yttervik, Servicekoordinator Norge. Tel: TELEPHONENO , e-post: EMAIL \n\nSend søknad merket "Servicetekniker Norge" på e-post til: EMAIL \n\nDenne rekrutteringen skjer helt i Part Constructions regi, og vi ønsker kun direkte kontakt med personlig søkere. \n\nVi gleder oss til å høre fra deg!', 'annonstextFormaterad': '<p><strong>Er du en handy og serviceinnstilt person som ønsker en spennende stilling i en fremoverlent og voksende bedrift?</strong></p><p>Som servicetekniker vil du ha ansvar for service og hjemmebesøk for Part Construction AB. Du vil behandle innkommende forespørsler, utføre service oppdrag på byggeprosjekter og reklamasjonsarbeid. Arbeidet foregår hovedsakelig på Østlandet, men vi har også leveranser til andre deler av landet. Noe reising vil derfor måtte påberegnes.</p><p><br></p><p><strong>Om deg:</strong></p><ul><li>En allsidig praktisk fagarbeider. Du er kanskje flislegger, rørlegger, elektriker eller har en annen relevant utdanning</li><li>Snakker og skriver godt Norsk</li><li>Førerkort klasse B/BE</li><li>God evne til kundebehandling</li><li>Er selvstendig og strukturert</li><li>Grunnleggende data kunnskaper.</li><li>Kjennskap til NS 3420, NS 8409 og Bustadoppføringslova.</li></ul><p><br></p><p><strong>Part Construction AB kan tilby:</strong></p><ul><li>Et dynamisk og spennende arbeidsmiljø</li><li>Utfordrende og variert arbeidsoppgaver</li><li>Konkurransedyktig betingelser</li><li>Servicebil, mobil telefon</li><li>Gode forsikringer og pensjons ordninger</li></ul><p><br></p><p><strong>Om arbeidsgiveren </strong></p><p>Part Construction AB, en av Europas ledende produsenter av prefabrikkerte baderoms moduler. Part Construction AB tilbyr entreprenørfirmaer og eiendomsselskaper prefabrikkerte baderom til blant annet, hoteller, sykehjem og boliger. Bademodulene leveres komplett med fliser, gulv, møbler, inventar, helt etter kundens ønsker. rørleggerarbeid og elektriske installasjoner er også ferdig montert fra fabrikk, Part er et familieeid selskap fra Kalix, med ca. 300 ansatte og med en omsetning på ca. 600 millioner. Hovedkvarter og produksjon er i Kalix / Överkalix, Norrbotten. Part Construction AB er en del av konserngruppen PartGroup sammen med søsterselskapene PreBad AB, Altor Industri, Isolamin AB, Space Interior AB, og PCS Modulsystem.</p><p><br></p><p><strong>Sektor</strong></p><p>Privat</p><p><strong>Sted</strong></p><p>Vestbygata 55 2003 Lillestrøm</p><p><strong>Bransje</strong></p><p>Bygg og anlegg,</p><p><strong>Stillingsfunksjon</strong></p><p>Håndverker / Andre montører,</p><p>Håndverker,</p><p>Håndverker / Flislegger, rørlegger, elektriker</p><p> </p><p><strong><u>Søknad</u></strong></p><p><br></p><p><strong>Tiltredelse</strong></p><p>Etter avtale</p><p><strong>Varighet</strong></p><p>Fast, heltid (6 måneders prøvetid)</p><p><strong>Bosted</strong></p><p>Oslo området</p><p><strong>Søknadsfrist</strong></p><p>Snarest, og innen 10.03.2021</p><p><strong>Spørsmål </strong>om stillingen rettes til FIRSTNAME LASTNAME Yttervik, Servicekoordinator Norge. Tel: TELEPHONENO , e-post: EMAIL </p><p>Send søknad merket "Servicetekniker Norge" på e-post til: EMAIL </p><p>Denne rekrutteringen skjer helt i Part Constructions regi, og vi ønsker kun direkte kontakt med personlig søkere. </p><p>Vi gleder oss til å høre fra deg!</p><p id="oppenforalla">Öppen för alla\nVi fokuserar på din kompetens, inte dina övriga förutsättningar. Vi är öppna för att anpassa rollen eller arbetsplatsen efter dina behov.</p>', 'ansokningssattEpost': ' EMAIL ', 'ansokningssattViaAF': False, 'anstallningTyp': {'namn': 'Vanlig anställning', 'varde': 'PFZr_Syz_cUq'}, 'antalPlatser': 1, 'arbetsgivareId': '20610987', 'arbetsgivareNamn': 'Partbyggen i Kalix AB', 'arbetsplatsId': '86440232', 'arbetsplatsNamn': 'Part Construction', 'arbetsplatsadress': {'gatuadress': None, 'kommun': None, 'koordinatPrecision': None, 'lan': None, 'land': {'namn': 'Norge', 'varde': 'QJgN_Zge_BzJ'}, 'latitud': None, 'longitud': None, 'postnr': None, 'postort': None}, 'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'ejKravSvenskaEngelska': False, 'informationAnsokningssatt': None, 'ingenErfarenhetKravs': False, 'inkluderande': True, 'kallaTyp': 'VIA_ANNONSERA', 'kontaktpersoner': [ {'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se', 'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}], 'korkort': [{'namn': 'B', 'varde': 'VTK8_WRx_GcM'}], 'lonTyp': {'namn': 'Fast månads- vecko- eller timlön', 'varde': 'oG8G_9cW_nRf'}, 'organisationsnummer': '5564003357', 'publiceringsdatum': '2021-01-04 16:18:13', 'referens': 'Servicetekniker Norge', 'sistaAnsokningsdatum': '2021-03-31 23:59:59', 'sistaPubliceringsdatum': '2022-06-14 14:52:55', 'sprak': [{'namn': 'Norska', 'varde': 'pnjj_2JX_Fub', 'vikt': 10}], 'telefonnummer': None, 'tillgangTillEgenBil': False, 'uppdateradTid': 1609773493583, 'varaktighetTyp': {'namn': 'Tillsvidare', 'varde': 'a7uU_j21_mkL'}, 'version': '1.0', 'webbadress': 'partab.nu', 'yrkeserfarenheter': [ {'erfarenhetsniva': {'namn': '1-2 års erfarenhet', 'varde': 'LLnd_5GJ_4ju'}, 'namn': 'VVS-montör', 'varde': 'ZZTi_v6g_4cZ', 'vikt': 5}], 'yrkesroll': {'namn': 'VVS-montör', 'varde': 'ZZTi_v6g_4cZ'}}, {'annonsId': 24513844, 'annonsrubrik': 'Svensktalande Mötesbokare till Soliga Malaga, Spanien!', 'annonstext': 'Svensktalande B2B mötesbokare till soliga Benalmadena i Spanien!\n \nVi söker nu en mötesbokare som kan hjälpa vår partner att ytterligare stärka sin position på företagsmarknaden. Vi söker säljare som är bra med människor, som vill utvecklas vidare som säljare som har förmågan att stå på sig och ge lite extra för att nå resultat. Medarbetare med goda empatiska färdigheter och som kan socialisera och prata med många olika personligheter.\n \nI din roll kommer du att boka möten för vår klient per telefon. Du kommer att arbeta för den svenska marknaden och kontakta verksamheter för att kunna väcka intresse, och sätta upp ett möte för den potentiella nya klienten och säljteamet.\n \nOm dig\nDu kommunicerar väl i tal och skrift samt tycker om att arbeta med uppsökande försäljning mot nya kunder. Vidare drivs du av bra värderingar och arbetar för att alla ska sträva mot samma mål. Likaså har du förmågan att ge det lilla extra för att uppnå dina egna mål. Du har ett starkt driv att uppnå goda resultat och är lösningsorienterad med en god arbetsmoral. Därtill har du hög förmåga att förvärva ny kunskap samt har erfarenhet av och trivs med att ha telefonen som arbetsredskap. \n\nVår klient erbjuder\n-Säljutbildning och vägledning i att arbeta med ett välkänt och starkt varumärke\n-Möjlighet att vidareutveckla dig själv med fokus på kompetensutveckling. \n-En ambitiös och stabil arbetsgivare med långsiktiga mål. \n-Konkurrenskraftig lön. \n-En energisk och ambitiös arbetsmiljö där alla hjälper varandra.\n\nAndra detaljer:\n-Arbetstider: Måndag-fredag 08:00-17:00\n-Heltidsanställning\n-Kontor i Benalmadena, Spanien\n\nKompetenskrav: \n-Flytande svenska både i tal och skrift \n-Flytande engelska både i tal och skrift \n-God teknisk förståelse\n\nFlyttpaket: Återbetald flybjljett + fyra veckors boende vid start samt hjälp med NIE-nummer.\n\nBeskrivning av företaget:\nVår klient är en välkänd aktör på företagsmarknaden - detta är en mycket spännande och utvecklande position. Denna position är placerad i vackra och soliga Benalmadena i Spanien. \n\nVi vill inte bli kontaktade av rekryterings-/bemanningsföretag', 'annonstextFormaterad': 'Svensktalande B2B mötesbokare till soliga Benalmadena i Spanien!\n \nVi söker nu en mötesbokare som kan hjälpa vår partner att ytterligare stärka sin position på företagsmarknaden. Vi söker säljare som är bra med människor, som vill utvecklas vidare som säljare som har förmågan att stå på sig och ge lite extra för att nå resultat. Medarbetare med goda empatiska färdigheter och som kan socialisera och prata med många olika personligheter.\n \nI din roll kommer du att boka möten för vår klient per telefon. Du kommer att arbeta för den svenska marknaden och kontakta verksamheter för att kunna väcka intresse, och sätta upp ett möte för den potentiella nya klienten och säljteamet.\n \nOm dig\nDu kommunicerar väl i tal och skrift samt tycker om att arbeta med uppsökande försäljning mot nya kunder. Vidare drivs du av bra värderingar och arbetar för att alla ska sträva mot samma mål. Likaså har du förmågan att ge det lilla extra för att uppnå dina egna mål. Du har ett starkt driv att uppnå goda resultat och är lösningsorienterad med en god arbetsmoral. Därtill har du hög förmåga att förvärva ny kunskap samt har erfarenhet av och trivs med att ha telefonen som arbetsredskap. \n\nVår klient erbjuder\n-Säljutbildning och vägledning i att arbeta med ett välkänt och starkt varumärke\n-Möjlighet att vidareutveckla dig själv med fokus på kompetensutveckling. \n-En ambitiös och stabil arbetsgivare med långsiktiga mål. \n-Konkurrenskraftig lön. \n-En energisk och ambitiös arbetsmiljö där alla hjälper varandra.\n\nAndra detaljer:\n-Arbetstider: Måndag-fredag 08:00-17:00\n-Heltidsanställning\n-Kontor i Benalmadena, Spanien\n\nKompetenskrav: \n-Flytande svenska både i tal och skrift \n-Flytande engelska både i tal och skrift \n-God teknisk förståelse\n\nFlyttpaket: Återbetald flybjljett + fyra veckors boende vid start samt hjälp med NIE-nummer.\n\nBeskrivning av företaget:\nVår klient är en välkänd aktör på företagsmarknaden - detta är en mycket spännande och utvecklande position. Denna position är placerad i vackra och soliga Benalmadena i Spanien. \n\nVi vill inte bli kontaktade av rekryterings-/bemanningsföretag', 'ansokningssattEpost': ' EMAIL ', 'ansokningssattViaAF': False, 'anstallningTyp': {'namn': 'Arbete utomlands', 'varde': '9Wuo_2Yb_36E'}, 'antalPlatser': 8, 'arbetsgivareId': '0', 'arbetsgivareNamn': 'Nordic Jobs Worldwide AS', 'arbetsplatsId': '21052284', 'arbetsplatsNamn': 'Nordic Jobs Worldwide AS', 'arbetsplatsadress': {'gatuadress': None, 'kommun': None, 'koordinatPrecision': None, 'lan': None, 'land': {'namn': 'Spanien', 'varde': 'bN7k_4ka_YGQ'}, 'latitud': None, 'longitud': None, 'postnr': None, 'postort': None}, 'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'ejKravSvenskaEngelska': False, 'externtAnnonsId': '0021-045667', 'informationAnsokningssatt': None, 'ingenErfarenhetKravs': True, 'inkluderande': False, 'kallaTyp': 'VIA_AIS', 'kontaktpersoner': [ {'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se', 'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}], 'lonTyp': {'namn': 'Fast och rörlig lön', 'varde': 'asrX_9Df_ukn'}, 'lonebeskrivning': 'Fast lön + provision', 'publiceringsdatum': '2021-02-01 10:57:26', 'sistaAnsokningsdatum': '2021-04-27 23:59:59', 'sistaPubliceringsdatum': '2022-06-14 14:55:21', 'sprak': [{'namn': 'Svenska', 'varde': 'zSLA_vw2_FXN', 'vikt': 10}, {'namn': 'Engelska', 'varde': 'NVxJ_hLg_TYS', 'vikt': 10}], 'telefonnummer': None, 'tillgangTillEgenBil': False, 'uppdateradTid': 1612173446756, 'varaktighetTyp': {'namn': 'Tillsvidare', 'varde': 'a7uU_j21_mkL'}, 'version': '2.0', 'villkorsbeskrivning': 'Arbetstider: Måndag-fredag 08:00-17:00', 'yrkesroll': {'namn': 'Besöksbokare/Kundbokare', 'varde': 'w8eg_Ufq_B5X'}}, ] ads_no_address = [ {'arbetsplatsadress': {'gatuadress': None, 'kommun': None, 'koordinatPrecision': None, 'lan': None, 'land': {'namn': None, 'varde': None}, 'latitud': None, 'longitud': None, 'postnr': None, 'postort': None}}, ]
177.297619
3,199
0.738065
5,899
44,679
5.568062
0.196983
0.004932
0.006333
0.001827
0.899805
0.890702
0.885526
0.87688
0.867655
0.860592
0
0.028646
0.160075
44,679
252
3,200
177.297619
0.846617
0
0
0.309623
0
0.054393
0.858259
0.116697
0
0
0
0
0
1
0
false
0.016736
0.008368
0
0.008368
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
1
1
0
0
0
0
0
0
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
9
faebbcbc5afca41622c242d6955bb7d2f1d7c2ee
3,174
py
Python
kachery/parser_helpers.py
flatironinstitute/kachery
d1076f6e8e2df26d3440fdb89f366ec44a502b9b
[ "Apache-2.0" ]
8
2020-03-05T19:41:03.000Z
2021-11-19T04:40:10.000Z
kachery/parser_helpers.py
flatironinstitute/kachery
d1076f6e8e2df26d3440fdb89f366ec44a502b9b
[ "Apache-2.0" ]
8
2019-11-15T03:40:07.000Z
2020-09-08T22:14:07.000Z
kachery/parser_helpers.py
flatironinstitute/kachery
d1076f6e8e2df26d3440fdb89f366ec44a502b9b
[ "Apache-2.0" ]
2
2020-08-06T19:56:12.000Z
2021-09-23T01:05:24.000Z
import os import kachery as ka def _add_download_args(parser): parser.add_argument('--fr', '-f', help='Where to load from', required=False, default='') parser.add_argument('--remote-only', action='store_true', help='Whether to only load from remote (good for debugging)') parser.add_argument('--url', help='The URL of the kachery database server to download from when loading from remote (or use KACHERY_URL environment variable)', required=False, default=None) parser.add_argument('--channel', '-c', help='The channel of the kachery database server to download from when loading from remote (or use KACHERY_CHANNEL environment variable)', required=False, default=None) parser.add_argument('--password', '-p', help='The password of the kachery database server to download from when loading from remote (or use KACHERY_PASSWORD environment variable)', required=False, default=None) def _add_upload_args(parser): parser.add_argument('--to', '-t', help='Where to store to', required=False, default='') parser.add_argument('--remote-only', action='store_true', help='Whether to only store to remote (good for saving disk space)') parser.add_argument('--url', help='The URL of the kachery database server to upload to when storing to remote (or use KACHERY_URL environment variable)', required=False, default=None) parser.add_argument('--channel', '-c', help='The channel of the kachery database server to upload to when storing to remote (or use KACHERY_CHANNEL environment variable)', required=False, default=None) parser.add_argument('--password', '-p', help='The password of the kachery database server to upload to when storing to remote (or use KACHERY_PASSWORD environment variable)', required=False, default=None) def _set_download_config_from_parsed_args(args): fr = args.fr or None url = args.url or None channel = args.channel or None password = args.password or None remote_only = args.remote_only if fr is not None: if url is not None or channel is not None or password is not None: raise Exception('Cannot use --url or --channel or --password together with --fr') ka.set_config( fr=fr, from_remote_only = remote_only ) else: ka.set_config( fr=dict( url=url, channel=channel, password=password ), from_remote_only = remote_only ) def _set_upload_config_from_parsed_args(args): to = args.to or None url = args.url or None channel = args.channel or None password = args.password or None remote_only = args.remote_only if to is not None: if url is not None or channel is not None or password is not None: raise Exception('Cannot use --url or --channel or --password together with --to') ka.set_config( to=to, to_remote_only = remote_only ) else: ka.set_config( to=dict( url=url, channel=channel, password=password ), to_remote_only = remote_only )
49.59375
214
0.669502
442
3,174
4.680995
0.147059
0.067666
0.082165
0.057999
0.885452
0.81102
0.81102
0.772354
0.738521
0.738521
0
0
0.23724
3,174
64
215
49.59375
0.854606
0
0
0.482759
0
0.103448
0.357795
0
0
0
0
0
0
1
0.068966
false
0.172414
0.034483
0
0.103448
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
1
0
0
0
0
0
7
879d727eb5dcbb7eb73833b8f03577b34d9910bb
78,721
py
Python
src/symbols.py
praga2018/SAMoCAD
4d1361e38bb2610753a588ca1d5e7391fa4a865e
[ "Apache-2.0" ]
5
2015-05-12T03:58:29.000Z
2020-05-01T14:20:54.000Z
src/symbols.py
praga2018/SAMoCAD
4d1361e38bb2610753a588ca1d5e7391fa4a865e
[ "Apache-2.0" ]
1
2019-07-24T01:20:18.000Z
2019-07-24T01:20:18.000Z
src/symbols.py
sallecta/PyTkCad
113994559d8d56c02dc102cce53bc7828588fc6c
[ "Apache-2.0" ]
2
2019-03-13T23:59:00.000Z
2022-01-27T07:04:49.000Z
# -*- coding: utf-8; -*- import rotate_object zoomm = 0.8 zoomp = 1.0/0.8 def font(x, y, text, size, zoomOLD, s_s, w_text, anchor, font, angle): if font == 'Architectural': tt = Text_arch(x, y, text, size, zoomOLD, s_s, w_text, anchor, angle) else: tt = Text_TXT(x, y, text, size, zoomOLD, s_s, w_text, anchor, angle) return tt class Text:#Общий класс шрифта def __init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle): if not zoomOLD: self.sy = -size else: if zoomOLD>0: self.sy = -size * (zoomp**zoomOLD) else: zoomOLDx=zoomOLD*(-1) self.sy = -size * (zoomm**zoomOLDx) self.w_text = w_text #Ширина символа self.s_s = s_s#смещение символов в строке self.anchor = anchor#Привязка надписи к базовой точке self.sx=self.sy/4.0#Ширина символа = четверть высоты self.liter = -self.s_s#Позиция первого символа self.Ltext = (len(text) * self.s_s * 2.0 * self.sx) - self.sx #- self.sx#Длина готовой строки axy = tuple(anchor) if axy[0] == 's': self.nachTextY = y else:# axy[0] == 'n': self.nachTextY = y + self.sy if axy[1] == 'w': self.nachTextX = x elif axy[1] == 'e': self.nachTextX = x-self.Ltext else:# axy[1] == 'c': self.nachTextX = x-self.Ltext/2.0 self.nabor = []#Список координат символов текста self.snapLine = [[self.nachTextX,self.nachTextY, self.nachTextX+self.Ltext,self.nachTextY]]#Координаты линии привязки текста self.nabor.extend(self.snapLine)#Добавить координаты в список #Дальше работает конкретный шрифт class Text_arch(Text):#Ахитектурный шрифт def __init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle): Text.__init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle) liter_dict = { u'А':self.a, 'A':self.a, 'a':self.a, u'а':self.a, u'б':self.r6, u'Б':self.r6, '6':self.r6, u'В':self.b, u'в':self.b, 'B':self.b, 'b':self.b, '8':self.b, u'С':self.c, 'C':self.c, 'c':self.c, u'с':self.c, u'Г':self.rg, u'г':self.rg, u'Д':self.rd, u'д':self.rd, u'Е':self.e, u'е':self.e, u'Ё':self.e, u'ё':self.e, 'e':self.e, 'E':self.e, u'Ж':self.rj, u'ж':self.rj, u'З':self.r3, u'з':self.r3, '3':self.r3, u'И':self.ri, u'и':self.ri, u'Й':self.rikr, u'й':self.rikr, u'К':self.k, u'к':self.k, 'K':self.k, 'k':self.k, u'Л':self.rl, u'л':self.rl, u'М':self.m, u'м':self.m, 'M':self.m, 'm':self.m, u'Н':self.h, u'н':self.h, 'H':self.h, 'h':self.h, u'О':self.o, u'о':self.o, 'O':self.o, 'o':self.o, '0':self.o, u'П':self.rp, u'п':self.rp, u'Р':self.p, u'р':self.p, 'P':self.p, 'p':self.p, u'Т':self.t, u'т':self.t, 'T':self.t, 't':self.t, u'У':self.y, u'у':self.y, 'Y':self.y, 'y':self.y, u'Ф':self.rf, u'ф':self.rf, u'Х':self.x, u'х':self.x, 'X':self.x, 'x':self.x, u'Ч':self.rch, u'ч':self.rch, u'Ц':self.rc, u'ц':self.rc, u'Ш':self.rsh, u'ш':self.rsh, u'Щ':self.rcsh, u'щ':self.rcsh, u'Ь':self.rmz, u'ь':self.rmz, u'Ъ':self.rtz, u'ъ':self.rtz, u'Ы':self.rii, u'ы':self.rii, u'Э':self.rae, u'э':self.rae, u'Ю':self.ru, u'ю':self.ru, u'Я':self.rya, u'я':self.rya, 'D':self.d, 'd':self.d, 'I':self.i, 'i':self.i, 'J':self.j, 'j':self.j, 'F':self.f, 'f':self.f, 'G':self.g, 'g':self.g, 'L':self.L, 'l':self.L, 'N':self.n, 'n':self.n, 'Q':self.q, 'q':self.q, 'R':self.r, 'r':self.r, 'S':self.r5, 's':self.r5, 'U':self.u, 'u':self.u, 'V':self.v, 'v':self.v, 'W':self.w, 'w':self.w, 'Z':self.z, 'z':self.z, '1':self.r1, '2':self.r2, '4':self.r4, '5':self.r5, '7':self.r7, '9':self.r9, ',':self.comma, '.':self.point, ';':self.c_point, ':':self.p_point, '!':self.emark, '?':self.qmark, '(':self.lbkt, ')':self.rbkt, '*':self.star, "/":self.dr, '+':self.plus, '-':self.minus, '=':self.equal, '>':self.bigest, '<':self.smollest, u'№':self.num, ' ':self.space, } for i in text:#Перебрать символы строки self.liter += self.s_s #Передвинуть позицию на один символ try: r = liter_dict[i]() except KeyError: r = self.qmark() self.nabor.extend(r)#Добавить в список координаты символа if angle: self.nabor = rotate_object.rotate_lines(x, y, self.nabor, angle) ''' if i in (u'А', 'A', 'a', u'а'):#Если символ А r = self.a()#Вызвать функцию, возвращающую список координат линий символа А elif i in (u'б', u'Б', '6'): r = self.r6() elif i in (u'В', u'в', 'B', 'b', '8'): r = self.b() elif i in (u'С', u'с', 'C', 'c'): r = self.c() elif i in (u'Г', u'г'): r = self.rg() elif i in (u'Д', u'д'): r = self.rd() elif i in (u'Е', u'е', u'Ё', u'ё', 'E', 'e'): r = self.e() elif i in (u'Ж', u'ж'): r = self.rj() elif i in (u'З', u'з', '3'): r = self.r3() elif i in (u'И', u'и'): r = self.ri() elif i in (u'Й', u'й'): r = self.rikr() elif i in (u'К', u'к', 'K', 'k'): r = self.k() elif i in (u'Л', u'л'): r = self.rl() elif i in (u'М', u'м', 'M', 'm'): r = self.m() elif i in (u'Н', u'н', 'H', 'h'): r = self.h() elif i in (u'О', u'о', 'O', 'o', '0'): r = self.o() elif i in (u'П', u'п'): r = self.rp() elif i in (u'Р', u'р', 'P', 'p'): r = self.p() elif i in (u'Т', u'т', 'T', 't'): r = self.t() elif i in (u'У', u'у', 'Y', 'y'): r = self.y() elif i in (u'Ф', u'ф'): r = self.rf() elif i in (u'Х', u'х', 'X', 'x'): r = self.x() elif i in (u'Ч', u'ч'): r = self.rch() elif i in (u'Ц', u'ц'): r = self.rc() elif i in (u'Ш', u'ш'): r = self.rsh() elif i in (u'Щ', u'щ'): r = self.rcsh() elif i in (u'Ь', u'ь'): r = self.rmz() elif i in (u'Ъ', u'ъ'): r = self.rtz() elif i in (u'Ы', u'ы'): r = self.rii() elif i in (u'Э', u'э'): r = self.rae() elif i in (u'Ю', u'ю'): r = self.ru() elif i in (u'Я', u'я'): r = self.rya() #Английские буквы (не объявленные ранее) elif i in ('D', 'd'): r = self.d() elif i in ('I', 'i'): r = self.i() elif i in ('J', 'j'): r = self.j() elif i in ('F', 'f'): r = self.f() elif i in ('G', 'g'): r = self.g() elif i in ('L', 'l'): r = self.L() elif i in ('N', 'n'): r = self.n() elif i in ('Q', 'q'): r = self.q() elif i in ('R', 'r'): r = self.r() elif i in ('U', 'u'): r = self.u() elif i in ('V', 'v'): r = self.v() elif i in ('W', 'w'): r = self.w() elif i in ('Z', 'z'): r = self.z() #Цифры (которые не объявлены ранее) elif i == '1': r = self.r1() elif i == '2': r = self.r2() elif i == '4': r = self.r4() elif i in ('5', 's', 'S'): r = self.r5() elif i == '7': r = self.r7() elif i == '9': r = self.r9() #Знаки препинания elif i == ',': r = self.comma() elif i == '.': r = self.point() elif i == ';': r = self.c_point() elif i == ':': r = self.p_point() elif i == '!': r = self.emark() elif i == '?': r = self.qmark() elif i == '(': r = self.lbkt() elif i == ')': r = self.rbkt() elif i == '*': r = self.star() elif i == "/": r = self.dr() elif i == '+': r = self.plus() elif i == '-': r = self.minus() elif i == '=': r = self.equal() elif i == '>': r = self.bigest() elif i == '<': r = self.smollest() elif i == u'№': r = self.num() elif i == ' ': r = self.space() else: r = self.qmark() ''' def ns(self, w = 1): #Определяет начало рисования символа w *= self.w_text #Относительная ширина x=self.nachTextX+self.sx*self.liter*2 #Начало рисования y=self.nachTextY y2 = self.sy #Высота символа x2 = self.sx * w #Ширина символа return x,y,x2,y2 #Символы узкого архитектурного шрифта def a(self):#Символ А x,y,x2,y2 = self.ns()#Определяет нижнюю левую точку символа и его ширину (x2) и высоту (y2) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0.5,-1.0) , #линия / в А (Y координата п рограмме перевернута) (1.0,0,0.5,-1.0) , #линия \ (0.25,-0.5,0.75,-0.5) , #линия - ]] def r6(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-0.75), (0,0,1.0,0), (0,-0.75,1.0,-0.75), (0,-1.0,1.0,-1.0), ]] def b(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,0), (0,-0.75,1.0,-0.75), (0,-1.0,1.0,-1.0), ]] def c(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,0,1.0,0), (0,-1.0,1.0,-1.0), ]] def rd(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0.5,-1.0), (1.0,0,0.5,-1.0), (-1.0/10.0,0,11.0/10.0,0), (-1.0/10.0,0,-1.0/10.0,1.0/8.0), (11.0/10.0,0,11.0/10.0,1.0/8.0), ]] def rg(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,1.0,-1.0), ]] def e(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,0,1.0,0), (0,-0.75,1.0,-0.75), (0,-1.0,1.0,-1.0), ]] def r3(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (0,0,1.0,0), (0,-0.75,1.0,-0.75), (0,-1.0,1.0,-1.0), ]] def ri(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-0.75,1.0,-1.0), ]] def rikr(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-0.75,1.0,-1.0), (0.5,-16.0/15.0,1.0,-7.0/6.0), ]] def k(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,-0.25,1.0,-1.0), (0,-0.25,1.0,-0.25), (0,-0.25,1.0,0), ]] def rl(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,1.0/2,-1.0), (1.0,0,1.0/2,-1.0), ]] def m(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,0.5,-0.75), (0.5,-0.75,1.0,-1.0), ]] def h(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-0.75,1.0,-0.75), ]] def o(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,0), (0,-1.0,1.0,-1.0), ]] def rp(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,1.0,-1.0), ]] def p(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,-0.25,1.0,-1.0), (0,-0.25,1.0,-0.25), (0,-1.0,1.0,-1.0), ]] def t(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0), (0,-1.0,1.0,-1.0), ]] def y(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (0,-0.25,0,-1.0), (0,-0.25,1.0,-0.25), (0,0,1.0,0), ]] def rf(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.25,0,-0.75), (1.0,-0.25,1.0,-0.75), (0,-0.25,1.0,-0.25), (0,-0.75,1.0,-0.75), (0.5,0,0.5,-1.0), ]] def x(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,1.0,-1.0), (0,-1.0,1.0,0), ]] def rch(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (0,-0.25,0,-1.0), (0,-0.25,1.0,-0.25), ]] def rc(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,11.0/10.0,0), (11.0/10.0,0,11.0/10.0,1/8.0), ]] def rsh(self): x,y,x2,y2 = self.ns(w = 1.5) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,0), (0.5,0,0.5,-1.0), ]] def rcsh(self): x,y,x2,y2 = self.ns(w = 1.5) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,11.0/10.0,0), (11.0/10.0,0,11.0/10.0,1/8.0), (0.5,0,0.5,-1.0), ]] def rmz(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-0.75), (0,0,1.0,0), (0,-0.75,1.0,-0.75), ]] def rtz(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-0.75), (0,0,1.0,0), (0,-0.75,1.0,-0.75), (0,-1.0,-1.0/6.0,-1.0), ]] def rii(self): x,y,x2,y2 = self.ns(1.5) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,0.66,0), (0,-0.75,0.66,-0.75), (0.66,0,0.66,-0.75), ]] def rae(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (0,0,1.0,0), (0,-0.5,1.0,-0.5), (0,-1.0,1.0,-1.0), ]] def ru(self): x,y,x2,y2 = self.ns(1.5) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0.33,0,0.33,-1.0), (0.33,0,1.0,0), (0.33,-1.0,1.0,-1.0), (0,-0.5,0.33,-0.5), ]] def rya(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (0,-0.25,0,-1.0), (0,-0.25,1.0,-0.25), (0,0,1.0,-0.25), (0,-1.0,1.0,-1.0), ]] def rj(self): x,y,x2,y2 = self.ns(w = 1.5) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.25,0,-1.0), (1.0,-0.25,1.0,-1.0), (0,-0.25,1.0,-0.25), (0.5,0,0.5,-1.0), (0.5,-0.25,0,0), (0.5,-0.25,1.0,0), ]] #Английские буквы (не объявленные ранее) def d(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,0,0.66,0), (0,-1.0,0.66,-1.0), (0.66,0,1.0,-0.25), (0.66,-1.0,1.0,-0.75), (1.0,-0.75,1.0,-0.25), ]] def i(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0), (0,-1.0,1.0,-1.0), (0,0,1.0,0), ]] def j(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0), (0,-1.0,1.0,-1.0), (0,0,0.5,0), (0,0,0,-1.0/5.0), ]] def f(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-0.75,1.0,-0.75), (0,-1.0,1.0,-1.0), ]] def g(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,0,1.0,0), (0,-1.0,1.0,-1.0), (1.0,0,1.0,-1.0/6.0), (7.0/6.0,-1.0/6.0,5.0/6.0,-1.0/6.0), ]] def L(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,0,1.0,0), ]] def n(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,1.0,0), ]] def q(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,0), (0,-1.0,1.0,-1.0), (4.0/5.0,-1.0/10.0,6.0/5.0,1/10.0), ]] def r(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,-0.25,1.0,-1.0), (0,-0.25,1.0,-0.25), (0,-1.0,1.0,-1.0), (1.0,0,0,-0.25), ]] def u(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,0), ]] def v(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.5,0), (1.0,-1.0,0.5,0), ]] def w(self): x,y,x2,y2 = self.ns(1.5) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.25,0), (0.5,-1.0,0.25,0), (0.5,-1.0,0.75,0), (0.75,0,1.0,-1.0), ]] def z(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-1.0,0,0), (0,-1.0,1.0,-1.0), (0,0,1.0,0), ]] #Цифры (не заданные ранее) def r1(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (0,-0.75,1.0,-1.0), ]] def r2(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-7.0/8.0,0,0), (1.0,-1.0,1.0,-7.0/8.0), (0,-7.0/8.0,0,-1.0), (0,-1.0,1.0,-1.0), (0,0,1.0,0), ]] def r4(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (1.0,-1.0,0,-0.25), (0,-0.25,1.0,-0.25), ]] def r5(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.75,0,-1.0), (1.0,0,1.0,-0.75), (0,0,1.0,0), (0,-0.75,1.0,-0.75), (0,-1.0,1.0,-1.0), ]] def r7(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-1.0,0,0), (0,-7.0/8.0,0,-1.0), (0,-1.0,1.0,-1.0), ]] def r9(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (0,-0.25,0,-1.0), (0,-0.25,1.0,-0.25), (0,0,1.0,0), (0,-1.0,1.0,-1.0), ]] #Знаки def comma(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,-0.33,1/5.0), ]] def point(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,-1.0/16.0,0), ]] def c_point(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,-0.33,1/5.0), (0,-1.0,-1.0/16.0,-1.0), ]] def p_point(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,-1.0/16.0,0), (0,-1.0,-1.0/16.0,-1.0), ]] def emark(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,-1.0/16.0,0), (-1.0/32.0,-1.0/10.0,-1.0/32.0,-1.0), ]] def qmark(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,-1.0/16.0,0), (1.0,-7.0/8.0,0,-0.25), (0,-0.25,0,-1.0/8.0), (1.0,-1.0,1.0,-7.0/8.0), (0,-7.0/8.0,0,-1.0), (0,-1.0,1.0,-1.0), ]] def lbkt(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,0,-1.0/4), (1.0,-1.0,0,-0.75), (0,-0.75,0,-0.25), ]] def rbkt(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,1.0,-0.25), (0,-1.0,1.0,-0.75), (1.0,-0.75,1.0,-0.25), ]] def star(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,1.0,-0.75), (1.0,-1.0,0,-0.75), (0.5,-1.0,0.5,-0.75), ]] def dr(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-1.0,0,0), ]] def plus(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,-0.66,0.5,-0.33), (0,-0.5,1.0,-0.5), ]] def minus(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.5,1.0,-0.5), ]] def equal(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,1.0,-0.33), (0,-0.5,1.0,-0.5), ]] def bigest(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0/3.5,1.0,-0.33), (0,-1.0/3.5,1.0,-0.25), ]] def smollest(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,1.0,-1.0/3.5), (0,-0.25,1.0,-1.0/3.5), ]] def num(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,1.0,0), (7.0/6.0,-1.0,8.0/6.0,-1.0), (1.33,-1.0,1.33,-0.75), (1.33,-0.75,7.0/6.0,-0.75), (7.0/6.0,-0.75,7.0/6.0,-1.0), (1.33,-0.5,7.0/6.0,-0.5), ]] def space(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ ]] class Text_TXT(Text): def __init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle): Text.__init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle) liter_dict = { u'А':self.a, 'A':self.a, 'a':self.am, u'а':self.am, u'б':self.rbm, u'Б':self.rb, u'В':self.b, u'в':self.rvm, 'B':self.b, 'b':self.bm, u'С':self.c, 'C':self.c, 'c':self.cm, u'с':self.cm, u'Г':self.rg, u'г':self.rgm, u'Д':self.rd, u'д':self.rdm, u'Е':self.e, u'е':self.em, u'Ё':self.e, u'ё':self.em, 'e':self.em, 'E':self.e, u'Ж':self.rj, u'ж':self.rjm, u'З':self.rz, u'з':self.rzm, u'И':self.ri, u'и':self.rim, u'Й':self.rikr, u'й':self.rikrm, u'К':self.k, u'к':self.rkm, 'K':self.k, 'k':self.km, u'Л':self.rl, u'л':self.rlm, u'М':self.m, u'м':self.rmm, 'M':self.m, 'm':self.mm, u'Н':self.h, u'н':self.rnm, 'H':self.h, 'h':self.hm, u'О':self.ro, u'о':self.om, 'O':self.o, 'o':self.om, u'П':self.rp, u'п':self.rpm, u'Р':self.p, u'р':self.pm, 'P':self.p, 'p':self.pm, u'Т':self.t, u'т':self.rtm, 'T':self.t, 't':self.tm, u'У':self.ru, u'у':self.rum, u'Ф':self.rf, u'ф':self.rfm, u'Х':self.x, u'х':self.xm, 'X':self.x, 'x':self.xm, u'Ч':self.rtch, u'ч':self.rtchm, u'Ц':self.rtc, u'ц':self.rtcm, u'Ш':self.rsh, u'ш':self.rshm, u'Щ':self.rtsh, u'щ':self.rtshm, u'Ь':self.rmznak, u'ь':self.rmznakm, u'Ъ':self.rtznak, u'ъ':self.rtznakm, u'Ы':self.rii, u'ы':self.riim, u'Э':self.rae, u'э':self.raem, u'Ю':self.ryu, u'ю':self.ryum, u'Я':self.rya, u'я':self.ryam, 'D':self.d, 'd':self.dm, 'I':self.i, 'i':self.im, 'J':self.j, 'j':self.jm, 'K':self.k, 'k':self.km, 'F':self.f, 'f':self.fm, 'G':self.g, 'g':self.gm, 'L':self.l, 'l':self.lm, 'N':self.n, 'n':self.nm, 'Q':self.q, 'q':self.qm, 'R':self.r, 'r':self.rm, 'S':self.s, 's':self.sm, 'U':self.u, 'u':self.um, 'V':self.v, 'v':self.vm, 'W':self.w, 'w':self.wm, 'Y':self.y, 'y':self.ym, 'Z':self.z, 'z':self.zm, '1':self.n1, '2':self.n2, '3':self.rz, '4':self.n4, '5':self.n5, '6':self.n6, '7':self.n7, '8':self.n8, '9':self.n9, '0':self.n0, ',':self.comma, '.':self.point, ';':self.c_point, ':':self.p_point, '!':self.emark, '?':self.qmark, '(':self.lbkt, ')':self.rbkt, '*':self.star, "/":self.dr, '+':self.plus, '-':self.minus, '=':self.equal, '>':self.bigest, '<':self.smollest, u'№':self.num, ' ':self.space, } for i in text: self.liter+=self.s_s try: r = liter_dict[i]() except KeyError: r = self.qmark() self.nabor.extend(r) ''' if i in ('A', u'А'): r = self.a() elif i in ('a', u'а'): r = self.am() elif i == u'Б': r = self.rb() elif i == u'б': r = self.rbm() elif i in ('B', u'В'): r = self.b() elif i =='b': r = self.bm() elif i in ('C', u'С'): r = self.c() elif i in ('c', u'с'): r = self.cm() elif i =='D': r = self.d() elif i =='d': r = self.dm() elif i in ('E', u'Е'): r = self.e() elif i in ('e', u'е'): r = self.em() elif i =='F': r = self.f() elif i =='f': r = self.fm() elif i =='G': r = self.g() elif i =='g': r = self.gm() elif i in ('H', u'Н'): r = self.h() elif i =='h': r = self.hm() elif i =='I': r = self.i() elif i =='i': r = self.im() elif i =='J': r = self.j() elif i =='j': r = self.jm() elif i in ('K', u'К'): r = self.k() elif i =='k': r = self.km() elif i =='L': r = self.l() elif i =='l': r = self.lm() elif i in ('M', u'М'): r = self.m() elif i =='m': r = self.mm() elif i =='N': r = self.n() elif i =='n': r = self.nm() elif i =='O': r = self.o() elif i in ('o', u'о'): r = self.om() elif i in ('P', u'Р'): r = self.p() elif i in ('p', u'р'): r = self.pm() elif i =='Q': r = self.q() elif i =='q': r = self.qm() elif i =='R': r = self.r() elif i =='r': r = self.rm() elif i =='S': r = self.s() elif i =='s': r = self.sm() elif i in ('T', u'Т'): r = self.t() elif i =='t': r = self.tm() elif i =='U': r = self.u() elif i =='u': r = self.um() elif i =='V': r = self.v() elif i =='v': r = self.vm() elif i =='W': r = self.w() elif i =='w': r = self.wm() elif i in ('X', u'Х'): r = self.x() elif i in ('x', u'х'): r = self.xm() elif i =='Y': r = self.y() elif i =='y': r = self.ym() elif i =='Z': r = self.z() elif i =='z': r = self.zm() elif i == u'в': r = self.rvm() elif i ==u'Г': r = self.rg() elif i ==u'г': r = self.rgm() elif i ==u'Д': r = self.rd() elif i ==u'д': r = self.rdm() elif i ==u'Ж': r = self.rj() elif i ==u'ж': r = self.rjm() elif i in (u'З', '3'): r = self.rz() elif i == u'з': r = self.rzm() elif i ==u'И': r = self.ri() elif i ==u'и': r = self.rim() elif i ==u'Й': r = self.rikr() elif i ==u'й': r = self.rikrm() elif i ==u'к': r = self.rkm() elif i ==u'Л': r = self.rl() elif i ==u'л': r = self.rlm() elif i ==u'м': r = self.rmm() elif i ==u'н': r = self.rnm() elif i ==u'П': r = self.rp() elif i ==u'п': r = self.rpm() elif i ==u'О': r = self.ro() elif i ==u'т': r = self.rtm() elif i ==u'У': r = self.ru() elif i ==u'у': r = self.rum() elif i ==u'Ф': r = self.rf() elif i ==u'ф': r = self.rfm() elif i ==u'Ц': r = self.rtc() elif i ==u'ц': r = self.rtcm() elif i ==u'Ч': r = self.rtch() elif i ==u'ч': r = self.rtchm() elif i ==u'Ш': r = self.rsh() elif i ==u'ш': r = self.rshm() elif i ==u'Щ': r = self.rtsh() elif i ==u'щ': r = self.rtshm() elif i ==u'Ъ': r = self.rtznak() elif i ==u'ъ': r = self.rtznakm() elif i ==u'Ы': r = self.rii() elif i ==u'ы': r = self.riim() elif i ==u'Ь': r = self.rmznak() elif i ==u'ь': r = self.rmznakm() elif i ==u'Э': r = self.rae() elif i ==u'э': r = self.raem() elif i ==u'Ю': r = self.ryu() elif i ==u'ю': r = self.ryum() elif i ==u'Я': r = self.rya() elif i ==u'я': r = self.ryam() elif i =='1': r = self.n1() elif i =='2': r = self.n2() elif i =='4': r = self.n4() elif i =='5': r = self.n5() elif i =='6': r = self.n6() elif i =='7': r = self.n7() elif i =='8': r = self.n8() elif i =='9': r = self.n9() elif i =='0': r = self.n0() elif i == ' ': r = self.space() #Знаки препинания elif i == ',': r = self.comma() elif i == '.': r = self.point() elif i == ';': r = self.c_point() elif i == ':': r = self.p_point() elif i == '!': r = self.emark() elif i == '?': r = self.qmark() elif i == '(': r = self.lbkt() elif i == ')': r = self.rbkt() elif i == '*': r = self.star() elif i == "/": r = self.dr() elif i == '+': r = self.plus() elif i == '-': r = self.minus() elif i == '=': r = self.equal() elif i == '>': r = self.bigest() elif i == '<': r = self.smollest() elif i == u'№': r = self.num() else: r = self.qmark() ''' if angle: self.nabor = rotate_object.rotate_lines(x, y, self.nabor, angle) def ns(self, w = 1): #Определяет начало рисования символа w *= self.w_text x=self.nachTextX+self.sx*self.liter*2 #Начало рисования y=self.nachTextY y2 = self.sy x2 = self.sx * w return x,y,x2,y2 def a(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-0.33), (1.0,0,1.0,-0.33), (0,-0.33,1.0,-0.33), (0,-0.33,0.5,-1.0), (1.0,-0.33,0.5,-1.0), ]] def am(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,0,1.0,-1.0), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def b(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,0.66,-1.0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-4.0/6.0), (1.0,-4.0/6.0,0.66,-0.5), ]] def bm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-0.33,1.0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (0,0,0,-1.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def c(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (0.33,0,0.66,0), (0.33,0,0,-0.33), ((1/3.0)*2.0,0,1.0,-0.33), ]] def cm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,1.0,-1.0), (0.33,0,1.0,0), (0.33,0,0,-0.33), ]] def d(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-0.33,1.0,-0.66), (-0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (0,0,0,-1.0), (-0.33,0,0.66,0), (0.66,0,1.0,-0.33), ]] def dm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,0,1.0,-1.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def e(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,1.0,-1.0), (0,0,1.0,0), (0,-0.5,0.5,-0.5), ]] def em(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.5,0.66,-0.5), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.75), (0,-0.33,0,-0.66), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,-0.5,1.0,-0.75), ]] def f(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,1.0,-1.0), (0,-0.5,0.5,-0.5), ]] def fm(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (-0.33,-0.5,0.66,-0.5), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.75), (0,0,0,-0.66), ]] def g(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,1.0,-1.0), (0.33,0,1.0,0), (0.33,0,0,-0.33), (1.0,0,1.0,-0.5), (1.0,-0.5,0.75,-0.5), ]] def gm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,0.33,1.0,-0.66), (0.33,0,1.0,0), (0.33,0,0,-0.33), (0.66,0.66,1.0,0.33), (0.33,0.66,0.66,0.66), (0.33,0.66,0,0.33), ]] def h(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-0.75,1.0,-0.75), ]] def hm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (0,0,0,-1.33), (1.0,0,1.0,-0.66), ]] def i(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.33,0,0.66,0), (0.33,-1.0,0.66,-1.0), (0.5,0,0.5,-1.0), ]] def im(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0), (0.5,-1.33,0.5,-3.5/3.0), ]] def j(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-0.33,1.0,-1.0), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def jm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-0.33,1.0,-1.0), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), (1.0,-1.33,1.0,-3.5/3.0), ]] def k(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-0.5,0.33,-0.5), (0.33,-0.5,1.0,-1.0), (0.33,-0.5,1.0,0), ]] def km(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.33), (0,-0.5,0.33,-0.5), (0.33,-0.5,1.0,-1.0), (0.33,-0.5,1.0,0), ]] def l(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,0,1.0,0), ]] def lm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,-0.25,0.5,-1.0), (0.5,-0.25,0.75,0), ]] def m(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,0.5,-0.5), (0.5,-0.5,1.0,-1.0), ]] def mm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.75,0.25,-1.0), (0.25,-1.0,0.5,-0.75), (0.5,-0.75,0.75,-1.0), (0.75,-1.0,1.0,-0.75), (0,0,0,-1.0), (1.0,0,1.0,-0.75), (0.5,-0.75,0.5,-0.5), ]] def n(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,1.0,0), ]] def nm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (0,0,0,-1.0), (1.0,0,1.0,-0.66), ]] def o(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,0), (0,-1.0,1.0,-1.0), ]] def om(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,-0.66,1.0,-0.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def p(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,0.66,-1.0), (0,-0.5,0.66,-0.5), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-4.0/6.0), (1.0,-4.0/6.0,0.66,-0.5), ]] def pm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-0.33,1.0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (0,0.5,0,-1.0), (0,0,0.66,0), (0.66,0,1.0,-0.33), ]] def q(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,-0.66,1.0,-0.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), (0.66,-0.33,1.0,0), ]] def qm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,0.66,1.0,-1.0), (0.33,0,1.0,0), (0.33,0,0,-0.33), ]] def r(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,0.66,-1.0), (0,-0.5,0.66,-0.5), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-4.0/6.0), (1.0,-4.0/6.0,0.66,-0.5), (0.33,-0.5,1.0,0), ]] def rm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (0,0,0,-1.0), ]] def s(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-5.0/6.0,1.0/6.0,-1.0), (1.0/6.0,-1.0,5.0/6.0,-1.0), (5.0/6.0,-1.0,1.0,-5.0/6.0), (0,-5.0/6.0,1.0,-1.0/6.0), (1.0/6.0,0,5.0/6.0,0), (1.0/6.0,0,0,-1.0/6.0), (5.0/6.0,0,1.0,-1.0/6.0), ]] def sm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.75,0.33,-1.0), (0.33,-1.0,1.0,-1.0), (0,-0.75,0.33,-0.5), (0.33,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.25), (1.0,-0.25,0.66,0), (0.66,0,0,0), ]] def t(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0/2,0,1.0/2,-1.0), (0,-1.0,1.0,-1.0), ]] def tm(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,-0.25,0.5,-1.0), (0,-0.66,1.0,-0.66), (0.5,-0.25,0.75,0), (0.75,0,1.0,-0.25), ]] def u(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-1.0), (1.0,-0.33,1.0,-1.0), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def um(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-1.0), (0,-0.33,0.33,0), (0.33,0,0.66,0), (1.0,-0.33,0.66,0), (1.0,0,1.0,-1.0), ]] def v(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.5,0), (0.5,0,1.0, -1), ]] def vm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.5,0), (0.5,0,1.0, -1.0), ]] def w(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.25,0), (0.25,0,0.5, -1/2.0), (0.5, -1/2.0,0.75, 0), (0.75, 0,1.0, -1), ]] def wm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.25,0), (0.25,0,0.5, -1.0), (0.5, -1.0,0.75, 0), (0.75, 0,1.0, -1.0), ]] def x(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,1.0,0), (0,0,1.0,-1.0), ]] def xm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,1.0,0), (0,0,1.0,-1.0), ]] def y(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.5,-0.5), (1.0, -1,0.5,-0.5), (0.5,-0.5,0.5,0), ]] def ym(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.5,-0.5), (0,0,1.0,-1.0), ]] def z(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-1.0,0,0), (0,-1.0,1.0,-1.0), (0,0,1.0,0), ]] def zm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-1.0,0,0), (0,-1.0,1.0,-1.0), (0,0,1.0,0), ]] #Русские буквы (которые не встречались ранее) def rb(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,0.66,-1.0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), (0.66,-1.0,0.66,-5.0/6.0), ]] def rbm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,0.66,-1.0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), ]] def rvm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,0.66,-1.0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-4.0/6.0), (1.0,-4.0/6.0,0.66,-0.5), ]] def rg(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,1.0,-1.0), (1.0,-1.0,1.0,-5.0/6.0), ]] def rgm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,1.0,-1.0), ]] def rd(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (-1.0/6.0,0,7.0/6.0,0), (-1.0/6.0,0,-1.0/6.0,1/6.0), (7.0/6.0,1/6.0,7.0/6.0,0), (0,0,0,-0.5), (0,-0.5,0.33,-1.0), (0.33,-1.0,1.0,-1.0), (1.0,-1.0,1.0,0), ]] def rdm(self): x,y,x2,y2 = self.ns(0.8) y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (-1.0/6.0,0,7.0/6.0,0), (-1.0/6.0,0,-1.0/6.0,1.0/6.0), (7.0/6.0,1.0/6.0,7.0/6.0,0), (0,0,0,-0.5), (0,-0.5,0.33,-1.0), (0.33,-1.0,1.0,-1.0), (1.0,-1.0,1.0,0), ]] def rj(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,1.0,-1.0), (0,-1.0,1.0,0), (0.5,0,0.5,-1.0), ]] def rjm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,1.0,-1.0), (0,-1.0,1.0,0), (0.5,0,0.5,-1.0), ]] def rz(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,-0.5,0.66,-0.5), (0.33,-1.0,0.66,-1.0), (0.33,0,0.66,0), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-4.0/6.0), (1.0,-4.0/6.0,0.66,-0.5), (0.33,-1.0,0,-5.0/6.0), (0.33,0,0,-1.0/6.0), ]] def rzm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,-0.5,0.66,-0.5), (0.33,-1.0,0.66,-1.0), (0.33,0,0.66,0), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-4.0/6.0), (1.0,-4.0/6.0,0.66,-0.5), (0.33,-1.0,0,-5.0/6.0), (0.33,0,0,-1.0/6.0), ]] def ri(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,-0.66), ]] def rim(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,-0.66), ]] def rikr(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,-0.66), (0.33,-1.0,0.66,-1.0), ]] def rikrm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,1.0,-0.66), (0.33,-1.0,0.66,-1.0), ]] def rkm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-0.5,0.33,-0.5), (0.33,-0.5,1.0,-1.0), (0.33,-0.5,1.0,0), ]] def rl(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.25,0,0.75,-1.0), (0.75,-1.0,1.0,-1.0), (1.0,0,1.0,-1.0), (0,0,0.25,0), ]] def rlm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0.5,-1.0), (0.5,-1.0,1.0,-1.0), (1.0,0,1.0,-1.0), ]] def rmm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,0.5,-0.5), (0.5,-0.5,1.0,-1.0), ]] def rnm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.+0.25,1.0,-1.+0.25), ]] def ro(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,-0.66,1.0,-0.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def rp(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,1.0,-1.0), ]] def rpm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,1.0,-1.0), ]] def rtm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0/2,0,1.0/2,-1.0), (0,-1.0,1.0,-1.0), ]] def ru(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0,-0.66), (0,-0.66,0.25,-0.5), (1.0,-1.0,1.0,-0.25), (1.0,-0.25,0.75,0), (0.75,0,0.25,0), (0.25,0,0,-0.25), (0.25,-0.5,1.0,-0.5), ]] def rum(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0.5,-0.5), (1.0,-1.0,1.0,-0.25), (1.0,-0.25,0.75,0), (0.75,0,0.25,0), (0.25,0,0,-0.25), (0.5,-0.5,1.0,-0.5), ]] def rf(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,-1.0,0.5,0), (0.25,-1.0,0.75,-1.0), (0.75,-1.0,1.0,-0.75), (1.0,-0.75,1.0,-0.5), (1.0,-0.5,0.75,-0.25), (0.75,-0.25,0.25,-0.25), (0.25,-0.25,0,-0.5), (0,-0.5,0,-0.75), (0,-0.75,0.25,-1.0), ]] def rfm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,-1.0,0.5,0), (0.25,-1.0,0.75,-1.0), (0.75,-1.0,1.0,-0.75), (1.0,-0.75,1.0,-0.5), (1.0,-0.5,0.75,-0.25), (0.75,-0.25,0.25,-0.25), (0.25,-0.25,0,-0.5), (0,-0.5,0,-0.75), (0,-0.75,0.25,-1.0), ]] def rtc(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,7.0/6.0,0), (7.0/6.0,0,7.0/6.0,1.0/6.0), ]] def rtcm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,0,7.0/6.0,0), (7.0/6.0,0,7.0/6.0,1.0/6.0), ]] def rtch(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0,-0.66), (0,-0.66,0.25,-0.5), (1.0,0,1.0,-1.0), (0.25,-0.5,1.0,-0.5), ]] def rtchm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,0,-0.5), (1.0,0,1.0,-1.0), (0,-0.5,1.0,-0.5), ]] def rsh(self): x,y,x2,y2 = self.ns(1.2) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0.5,-0.5,0.5,0), (0,0,1.0,0), ]] def rshm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0.5,-0.5,0.5,0), (0,0,1.0,0), ]] def rtsh(self): x,y,x2,y2 = self.ns(1.2) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0.5,-0.5,0.5,0), (0,0,7.0/6.0,0), (7.0/6.0,0,7.0/6.0,1.0/6.0), ]] def rtshm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0.5,-0.5,0.5,0), (0,0,7.0/6.0,0), (7.0/6.0,0,7.0/6.0,1.0/6.0), ]] def rtznak(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,-1.0/6.0,-1.0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), (-1.0/6.0,-1.0,-1.0/6.0,-5.0/6.0), ]] def rtznakm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,-1.0,-1.0/6.0,-1.0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), (-1.0/6.0,-1.0,-1.0/6.0,-5.0/6.0), ]] def rii(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (7.0/6.0,-1.0,7.0/6.0,0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), ]] def riim(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (7.0/6.0,-1.0,7.0/6.0,0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), ]] def rmznak(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), ]] def rmznakm(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (0,0,0.66,0), (0,-0.5,0.66,-0.5), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-1.0/6.0), (1.0,-1.0/6.0,0.66,0), ]] def rae(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.33,-0.5,1.0,-0.5), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,-0.66,1.0,-0.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def raem(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.33,-0.5,1.0,-0.5), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,-0.66,1.0,-0.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), ]] def ryu(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,-0.66,1.0,-0.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), (-1.0/6.0,-0.5,0,-0.5), (-1.0/6.0,0,-1.0/6.0,-1.0), ]] def ryum(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,0,-0.66), (0,-0.66,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-0.66), (1.0,-0.66,1.0,-0.33), (0.33,0,0.66,0), (0.33,0,0,-0.33), (0.66,0,1.0,-0.33), (-1.0/6.0,-0.5,0,-0.5), (-1.0/6.0,0,-1.0/6.0,-1.0), ]] def rya(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (1.0,-1.0,0.33,-1.0), (0.33,-1.0,0,-0.75), (0,-0.75,0,-0.5), (0,-0.5,0.25, -1/4.0), (0.25, -1/4.0,1.0, -1/4.0), (0.5, -1/4.0,0, 0), ]] def ryam(self): x,y,x2,y2 = self.ns() y2 *= 0.75 return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,1.0,-1.0), (1.0,-1.0,0.33,-1.0), (0.33,-1.0,0,-0.75), (0,-0.75,0,-0.5), (0,-0.5,0.25, -0.25), (0.25, -0.25,1.0, -0.25), (0.5, -0.25,0, 0), ]] #Цифры def n1(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.33,0,0.66,0), (0.5,-1.0,0.33,-5.0/6.0), (0.5,0,0.5,-1.0), ]] def n2(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-5.0/6.0,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-0.66), (1.0,-0.66,0.66,-0.5), (0.66,-0.5,0.33,-0.5), (0.33,-0.5,0,-0.33), (0,-0.33,0,0), (0,0,1.0,0), ]] def n4(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-1.0,1.0,0), (1.0,-1.0,0,-0.5), (0,-0.5,7.0/6.0,-0.5), ]] def n5(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-1.0,0,-1.0), (0,-1.0,0,-0.75), (0,-0.75,0.66,-0.75), (0.66,-0.75,1.0,-0.5), (1.0,-0.5,1.0,-0.25), (1.0,-0.25,0.66,0), (0.66,0,0.33,0), (0.33,0,0,-0.25), ]] def n6(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.66,-1.0,0.33,-1.0), (0.33,-1.0,0,-0.75), (0,-0.75,0,-0.25), (0.66,-0.5,1.0,-0.33), (1.0,-0.33,1.0,-0.25), (1.0,-0.25,0.66,0), (0.66,0,0.33,0), (0.33,0,0,-0.25), (0,-0.5,0.66,-0.5), ]] def n7(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0,1.0,-1.0), (1.0,-1.0,0,0), ]] def n8(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-5.0/6.0,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-0.66), (1.0,-0.66,0.66,-0.5), (0.66,-0.5,0.33,-0.5), (0.33,-0.5,0,-0.33), (0,-0.33,0,-1.0/6.0), (0,-1.0/6.0,0.33,0), (0.33,0,0.66,0), (0.66,0,1.0,-1.0/6.0), (1.0,-1.0/6.0,1.0,-0.33), (1.0,-0.33,0.66,-0.5), (0,-5.0/6.0,0,-0.66), (0,-0.66,0.33,-0.5), ]] def n9(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-5.0/6.0,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-1.0/6.0), (1.0,-0.5,0.33,-0.5), (0.33,0,0.66,0), (0.66,0,1.0,-1.0/6.0), (0,-5.0/6.0,0,-0.66), (0,-0.66,0.33,-0.5), ]] def n0(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-1.0/6.0,0,-5.0/6.0), (0,-5.0/6.0,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,1.0,-1.0/6.0), (0.33,0,0.66,0), (0.33,0,0,-1.0/6.0), (0.66,0,1.0,-1.0/6.0), ]] #Знаки препинания def point(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0/6.0), ]] def comma(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0/6.0), (0.5,0,0,0.33), ]] def c_point(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0/6.0), (0.5,0,0,0.33), (0.5,-0.33,0.5,-0.5), ]] def p_point(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0/6.0), (0.5,-0.33,0.5,-0.5), ]] def emark(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,0,0.5,-1.0/6.0), (0.5,-0.33,0.5,-1.0), ]] def space(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ ]] def qmark(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-5.0/6.0,0.33,-1.0), (0.33,-1.0,0.66,-1.0), (0.66,-1.0,1.0,-5.0/6.0), (1.0,-5.0/6.0,0.5,-0.5), (0.5,0,0.5,-1.0/6.0), (0.5,-0.33,0.5,-0.5), ]] def lbkt(self): x,y,x2,y2 = self.ns(0.5) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,0,0,-0.33), (0,-0.33,0,-0.66), (0,-0.66,1.0,-1.0), ]] def rbkt(self): x,y,x2,y2 = self.ns(0.5) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,1.0,-0.33), (1.0,-0.33,1.0,-0.66), (1.0,-0.66,0,-1.0), ]] def star(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,1.0,-0.66), (1.0,-0.33,0,-0.66), (0,-0.5,1.0,-0.5), ]] def dr(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,1.0,-1.0), ]] def plus(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0.5,-0.33,0.5,-0.66), (0,-0.5,1.0,-0.5), ]] def minus(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.5,1.0,-0.5), ]] def equal(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.33,1.0,-0.33), (0,-0.66,1.0,-0.66), ]] def smollest(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,-0.5,1.0,-0.33), (0,-0.5,1.0,-0.66), ]] def bigest(self): x,y,x2,y2 = self.ns(0.8) return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (1.0,-0.5,0,-0.33), (1.0,-0.5,0,-0.66), ]] def num(self): x,y,x2,y2 = self.ns() return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [ (0,0,0,-1.0), (1.0,0,1.0,-1.0), (0,-1.0,1.0,0), (7.0/6.0,-1.0,1.33,-1.0), (1.33,-1.0,1.33,-0.75), (1.33,-0.75,7.0/6.0,-0.75), (7.0/6.0,-0.75,7.0/6.0,-1.0), (1.33,-0.5,7.0/6.0,-0.5), ]]
27.984714
132
0.342018
15,001
78,721
1.790614
0.021065
0.102007
0.085552
0.051971
0.888649
0.836938
0.806001
0.765087
0.755668
0.738357
0
0.18821
0.406321
78,721
2,812
133
27.994666
0.38637
0.011725
0
0.798845
0
0
0.004709
0
0
0
0
0
0
1
0.096246
false
0
0.000481
0
0.192974
0
0
0
1
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
1
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
9
87a47ac74561dc50468225283ac64ebcb6a54698
32,272
py
Python
mailslurp_client/api/domain_controller_api.py
mailslurp/mailslurp-client-python
a1e9fdc6eb06e192909fd57a64813beb32419594
[ "MIT" ]
6
2020-04-30T07:47:42.000Z
2022-03-24T20:58:58.000Z
mailslurp_client/api/domain_controller_api.py
mailslurp/mailslurp-client-python
a1e9fdc6eb06e192909fd57a64813beb32419594
[ "MIT" ]
1
2020-09-20T19:58:21.000Z
2020-11-29T16:49:19.000Z
mailslurp_client/api/domain_controller_api.py
mailslurp/mailslurp-client-python
a1e9fdc6eb06e192909fd57a64813beb32419594
[ "MIT" ]
1
2019-08-09T14:55:50.000Z
2019-08-09T14:55:50.000Z
# coding: utf-8 """ MailSlurp API MailSlurp is an API for sending and receiving emails from dynamically allocated email addresses. It's designed for developers and QA teams to test applications, process inbound emails, send templated notifications, attachments, and more. ## Resources - [Homepage](https://www.mailslurp.com) - Get an [API KEY](https://app.mailslurp.com/sign-up/) - Generated [SDK Clients](https://www.mailslurp.com/docs/) - [Examples](https://github.com/mailslurp/examples) repository # noqa: E501 The version of the OpenAPI document: 6.5.2 Generated by: https://openapi-generator.tech """ from __future__ import absolute_import import re # noqa: F401 # python 2 and python 3 compatibility library import six from mailslurp_client.api_client import ApiClient from mailslurp_client.exceptions import ( # noqa: F401 ApiTypeError, ApiValueError ) class DomainControllerApi(object): """NOTE: This class is auto generated by OpenAPI Generator Ref: https://openapi-generator.tech Do not edit the class manually. """ def __init__(self, api_client=None): if api_client is None: api_client = ApiClient() self.api_client = api_client def add_domain_wildcard_catch_all(self, id, **kwargs): # noqa: E501 """Add catch all wild card inbox to domain # noqa: E501 Add a catch all inbox to a domain so that any emails sent to it that cannot be matched will be sent to the catch all inbox generated # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.add_domain_wildcard_catch_all(id, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str id: id (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: DomainDto If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True return self.add_domain_wildcard_catch_all_with_http_info(id, **kwargs) # noqa: E501 def add_domain_wildcard_catch_all_with_http_info(self, id, **kwargs): # noqa: E501 """Add catch all wild card inbox to domain # noqa: E501 Add a catch all inbox to a domain so that any emails sent to it that cannot be matched will be sent to the catch all inbox generated # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.add_domain_wildcard_catch_all_with_http_info(id, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str id: id (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(DomainDto, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = [ 'id' ] all_params.extend( [ 'async_req', '_return_http_data_only', '_preload_content', '_request_timeout' ] ) for key, val in six.iteritems(local_var_params['kwargs']): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method add_domain_wildcard_catch_all" % key ) local_var_params[key] = val del local_var_params['kwargs'] # verify the required parameter 'id' is set if self.api_client.client_side_validation and ('id' not in local_var_params or # noqa: E501 local_var_params['id'] is None): # noqa: E501 raise ApiValueError("Missing the required parameter `id` when calling `add_domain_wildcard_catch_all`") # noqa: E501 collection_formats = {} path_params = {} if 'id' in local_var_params: path_params['id'] = local_var_params['id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['API_KEY'] # noqa: E501 return self.api_client.call_api( '/domains/{id}/wildcard', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='DomainDto', # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get('async_req'), _return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501 _preload_content=local_var_params.get('_preload_content', True), _request_timeout=local_var_params.get('_request_timeout'), collection_formats=collection_formats) def create_domain(self, domain_options, **kwargs): # noqa: E501 """Create Domain # noqa: E501 Link a domain that you own with MailSlurp so you can create email addresses using it. Endpoint returns DNS records used for validation. You must add these verification records to your host provider's DNS setup to verify the domain. # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_domain(domain_options, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param CreateDomainOptions domain_options: domainOptions (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: DomainDto If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True return self.create_domain_with_http_info(domain_options, **kwargs) # noqa: E501 def create_domain_with_http_info(self, domain_options, **kwargs): # noqa: E501 """Create Domain # noqa: E501 Link a domain that you own with MailSlurp so you can create email addresses using it. Endpoint returns DNS records used for validation. You must add these verification records to your host provider's DNS setup to verify the domain. # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_domain_with_http_info(domain_options, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param CreateDomainOptions domain_options: domainOptions (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(DomainDto, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = [ 'domain_options' ] all_params.extend( [ 'async_req', '_return_http_data_only', '_preload_content', '_request_timeout' ] ) for key, val in six.iteritems(local_var_params['kwargs']): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method create_domain" % key ) local_var_params[key] = val del local_var_params['kwargs'] # verify the required parameter 'domain_options' is set if self.api_client.client_side_validation and ('domain_options' not in local_var_params or # noqa: E501 local_var_params['domain_options'] is None): # noqa: E501 raise ApiValueError("Missing the required parameter `domain_options` when calling `create_domain`") # noqa: E501 collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if 'domain_options' in local_var_params: body_params = local_var_params['domain_options'] # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['application/json']) # noqa: E501 # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.select_header_content_type( # noqa: E501 ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['API_KEY'] # noqa: E501 return self.api_client.call_api( '/domains', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='DomainDto', # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get('async_req'), _return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501 _preload_content=local_var_params.get('_preload_content', True), _request_timeout=local_var_params.get('_request_timeout'), collection_formats=collection_formats) def delete_domain(self, id, **kwargs): # noqa: E501 """Delete a domain # noqa: E501 Delete a domain. This will disable any existing inboxes that use this domain. # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_domain(id, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str id: id (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: list[str] If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True return self.delete_domain_with_http_info(id, **kwargs) # noqa: E501 def delete_domain_with_http_info(self, id, **kwargs): # noqa: E501 """Delete a domain # noqa: E501 Delete a domain. This will disable any existing inboxes that use this domain. # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_domain_with_http_info(id, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str id: id (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(list[str], status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = [ 'id' ] all_params.extend( [ 'async_req', '_return_http_data_only', '_preload_content', '_request_timeout' ] ) for key, val in six.iteritems(local_var_params['kwargs']): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method delete_domain" % key ) local_var_params[key] = val del local_var_params['kwargs'] # verify the required parameter 'id' is set if self.api_client.client_side_validation and ('id' not in local_var_params or # noqa: E501 local_var_params['id'] is None): # noqa: E501 raise ApiValueError("Missing the required parameter `id` when calling `delete_domain`") # noqa: E501 collection_formats = {} path_params = {} if 'id' in local_var_params: path_params['id'] = local_var_params['id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['API_KEY'] # noqa: E501 return self.api_client.call_api( '/domains/{id}', 'DELETE', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='list[str]', # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get('async_req'), _return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501 _preload_content=local_var_params.get('_preload_content', True), _request_timeout=local_var_params.get('_request_timeout'), collection_formats=collection_formats) def get_domain(self, id, **kwargs): # noqa: E501 """Get a domain # noqa: E501 Returns domain verification status and tokens for a given domain # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_domain(id, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str id: id (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: DomainDto If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True return self.get_domain_with_http_info(id, **kwargs) # noqa: E501 def get_domain_with_http_info(self, id, **kwargs): # noqa: E501 """Get a domain # noqa: E501 Returns domain verification status and tokens for a given domain # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_domain_with_http_info(id, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str id: id (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(DomainDto, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = [ 'id' ] all_params.extend( [ 'async_req', '_return_http_data_only', '_preload_content', '_request_timeout' ] ) for key, val in six.iteritems(local_var_params['kwargs']): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method get_domain" % key ) local_var_params[key] = val del local_var_params['kwargs'] # verify the required parameter 'id' is set if self.api_client.client_side_validation and ('id' not in local_var_params or # noqa: E501 local_var_params['id'] is None): # noqa: E501 raise ApiValueError("Missing the required parameter `id` when calling `get_domain`") # noqa: E501 collection_formats = {} path_params = {} if 'id' in local_var_params: path_params['id'] = local_var_params['id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['API_KEY'] # noqa: E501 return self.api_client.call_api( '/domains/{id}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='DomainDto', # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get('async_req'), _return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501 _preload_content=local_var_params.get('_preload_content', True), _request_timeout=local_var_params.get('_request_timeout'), collection_formats=collection_formats) def get_domains(self, **kwargs): # noqa: E501 """Get domains # noqa: E501 List all custom domains you have created # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_domains(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: list[DomainPreview] If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True return self.get_domains_with_http_info(**kwargs) # noqa: E501 def get_domains_with_http_info(self, **kwargs): # noqa: E501 """Get domains # noqa: E501 List all custom domains you have created # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_domains_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(list[DomainPreview], status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = [ ] all_params.extend( [ 'async_req', '_return_http_data_only', '_preload_content', '_request_timeout' ] ) for key, val in six.iteritems(local_var_params['kwargs']): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method get_domains" % key ) local_var_params[key] = val del local_var_params['kwargs'] collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['API_KEY'] # noqa: E501 return self.api_client.call_api( '/domains', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='list[DomainPreview]', # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get('async_req'), _return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501 _preload_content=local_var_params.get('_preload_content', True), _request_timeout=local_var_params.get('_request_timeout'), collection_formats=collection_formats) def update_domain(self, id, update_domain_dto, **kwargs): # noqa: E501 """Update a domain # noqa: E501 Update values on a domain. Note you cannot change the domain name as it is immutable. Recreate the domain if you need to alter this. # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_domain(id, update_domain_dto, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str id: id (required) :param UpdateDomainOptions update_domain_dto: updateDomainDto (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: DomainDto If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True return self.update_domain_with_http_info(id, update_domain_dto, **kwargs) # noqa: E501 def update_domain_with_http_info(self, id, update_domain_dto, **kwargs): # noqa: E501 """Update a domain # noqa: E501 Update values on a domain. Note you cannot change the domain name as it is immutable. Recreate the domain if you need to alter this. # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_domain_with_http_info(id, update_domain_dto, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str id: id (required) :param UpdateDomainOptions update_domain_dto: updateDomainDto (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(DomainDto, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = [ 'id', 'update_domain_dto' ] all_params.extend( [ 'async_req', '_return_http_data_only', '_preload_content', '_request_timeout' ] ) for key, val in six.iteritems(local_var_params['kwargs']): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method update_domain" % key ) local_var_params[key] = val del local_var_params['kwargs'] # verify the required parameter 'id' is set if self.api_client.client_side_validation and ('id' not in local_var_params or # noqa: E501 local_var_params['id'] is None): # noqa: E501 raise ApiValueError("Missing the required parameter `id` when calling `update_domain`") # noqa: E501 # verify the required parameter 'update_domain_dto' is set if self.api_client.client_side_validation and ('update_domain_dto' not in local_var_params or # noqa: E501 local_var_params['update_domain_dto'] is None): # noqa: E501 raise ApiValueError("Missing the required parameter `update_domain_dto` when calling `update_domain`") # noqa: E501 collection_formats = {} path_params = {} if 'id' in local_var_params: path_params['id'] = local_var_params['id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if 'update_domain_dto' in local_var_params: body_params = local_var_params['update_domain_dto'] # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['application/json']) # noqa: E501 # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.select_header_content_type( # noqa: E501 ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['API_KEY'] # noqa: E501 return self.api_client.call_api( '/domains/{id}', 'PUT', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='DomainDto', # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get('async_req'), _return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501 _preload_content=local_var_params.get('_preload_content', True), _request_timeout=local_var_params.get('_request_timeout'), collection_formats=collection_formats)
44.208219
487
0.589768
3,621
32,272
5.022646
0.071251
0.040908
0.055424
0.029692
0.938143
0.925771
0.915819
0.911585
0.908616
0.89311
0
0.014193
0.338498
32,272
729
488
44.268861
0.837737
0.477132
0
0.730205
0
0
0.15965
0.032199
0
0
0
0
0
1
0.038123
false
0
0.014663
0
0.090909
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
87eba730d3fc393d69e8c1f01825615681524d41
2,563
py
Python
payment/filter.py
AhmedElmawary/erp
c998787c62194e26e10e3cbc61e35935e901e56d
[ "MIT" ]
null
null
null
payment/filter.py
AhmedElmawary/erp
c998787c62194e26e10e3cbc61e35935e901e56d
[ "MIT" ]
null
null
null
payment/filter.py
AhmedElmawary/erp
c998787c62194e26e10e3cbc61e35935e901e56d
[ "MIT" ]
null
null
null
from django.contrib import admin from django.utils.translation import gettext_lazy from payment.models import PaymentTransactionType class ClientTransactionType(admin.SimpleListFilter): # Human-readable title which will be displayed in the # right admin sidebar just above the filter options. title = gettext_lazy('Transaction type') # Parameter for the filter that will be used in the URL query. # parameter_name = 'type_tranasction__id__exact' parameter_name = 'transactions_type' def lookups(self, request, model_admin): """ Returns a list of tuples. The first element in each tuple is the coded value for the option that will appear in the URL query. The second element is the human-readable name for the option that will appear in the right sidebar. """ options_for_client = [] for row in PaymentTransactionType.objects.filter(for_client=True): options_for_client.append((row.id, str(row))) return options_for_client def queryset(self, request, queryset): """ Returns the filtered queryset based on the value provided in the query string and retrievable via `self.value()`. """ if self.value(): return queryset.filter(type_tranasction=self.value()) class SupplierTransactionType(admin.SimpleListFilter): # Human-readable title which will be displayed in the # right admin sidebar just above the filter options. title = gettext_lazy('Transaction type') # Parameter for the filter that will be used in the URL query. # parameter_name = 'type_tranasction__id__exact' parameter_name = 'transactions_type' def lookups(self, request, model_admin): """ Returns a list of tuples. The first element in each tuple is the coded value for the option that will appear in the URL query. The second element is the human-readable name for the option that will appear in the right sidebar. """ options_for_supplier = [] for row in PaymentTransactionType.objects.filter(for_supplier=True): options_for_supplier.append((row.id, str(row))) return options_for_supplier def queryset(self, request, queryset): """ Returns the filtered queryset based on the value provided in the query string and retrievable via `self.value()`. """ if self.value(): return queryset.filter(type_tranasction=self.value())
37.691176
76
0.680062
325
2,563
5.249231
0.252308
0.029308
0.023447
0.030481
0.851114
0.851114
0.851114
0.797186
0.758499
0.758499
0
0
0.253609
2,563
67
77
38.253731
0.891793
0.430745
0
0.48
0
0
0.052257
0
0
0
0
0
0
1
0.16
false
0
0.12
0
0.68
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
1
0
0
7
3576c087b9c0cfbb0f42b8bc708af9a24746ec4d
14,707
py
Python
ScriptArchive/LSTM_multi_detect_FranklinBasin.py
AmberSJones/PyHydroQC
9f8992672ce3163eb048964e85680c526b4fd3f3
[ "BSD-3-Clause" ]
6
2021-04-22T15:12:55.000Z
2022-02-01T00:47:15.000Z
ScriptArchive/LSTM_multi_detect_FranklinBasin.py
AmberSJones/PyHydroQC
9f8992672ce3163eb048964e85680c526b4fd3f3
[ "BSD-3-Clause" ]
6
2020-12-22T14:12:30.000Z
2021-04-02T22:53:47.000Z
ScriptArchive/LSTM_multi_detect_FranklinBasin.py
AmberSJones/PyHydroQC
9f8992672ce3163eb048964e85680c526b4fd3f3
[ "BSD-3-Clause" ]
6
2021-01-06T08:09:59.000Z
2022-02-01T00:46:22.000Z
################################ # LSTM DEVELOP AND DIAGNOSTIC # ################################ # This code takes raw data and corrected data, applies an LSTM model, and identifies anomalies. import rules_detect import anomaly_utilities import modeling_utilities import numpy as np import tensorflow as tf import pandas as pd import seaborn as sns from matplotlib.pylab import rcParams import matplotlib.pyplot as plt import plotly.io as pio pio.renderers.default = "browser" pd.options.mode.chained_assignment = None sns.set(style='whitegrid', palette='muted') rcParams['figure.figsize'] = 14, 8 np.random.seed(1) print('Tensorflow version:', tf.__version__) print("LSTM exploration script begin.") ################################################## # LSTM Multivariate Retrieve and Preprocess Data # ################################################## # DEFINE SITE and VARIABLE # ######################################### # site = "BlackSmithFork" site = "FranklinBasin" # site = "MainStreet" # site = "Mendon" # site = "TonyGrove" # site = "WaterLab" sensor = ['temp', 'cond', 'ph', 'do'] year = [2014, 2015, 2016, 2017, 2018, 2019] # GET DATA # ######################################### df_full, sensor_array = anomaly_utilities.get_data(site, sensor, year, path="/LRO_data/") # RULES BASED DETECTION # ######################################### maximum = [13, 380, 9.2, 13] minimum = [-2, 120, 7.5, 8] length = 6 size = [] for i in range(0, len(sensor_array)): sensor_array[sensor[i]] = rules_detect.range_check(sensor_array[sensor[i]], maximum[i], minimum[i]) sensor_array[sensor[i]] = rules_detect.persistence(sensor_array[sensor[i]], length) s = rules_detect.group_size(sensor_array[sensor[i]]) size.append(s) sensor_array[sensor[i]] = rules_detect.interpolate(sensor_array[sensor[i]]) # Create new data frame with raw and corrected data for variables of interest df_observed = pd.DataFrame(index=df_full.index) df_observed['temp_obs'] = sensor_array['temp']['observed'] df_observed['cond_obs'] = sensor_array['cond']['observed'] df_observed['ph_obs'] = sensor_array['ph']['observed'] df_observed['do_obs'] = sensor_array['do']['observed'] df_raw = pd.DataFrame(index=df_full.index) df_raw['temp'] = df_full['temp'] df_raw['cond'] = df_full['cond'] df_raw['ph'] = df_full['ph'] df_raw['do'] = df_full['do'] df_anomaly = pd.DataFrame(index=df_full.index) df_anomaly['temp_anom'] = sensor_array['temp']['anomaly'] df_anomaly['cond_anom'] = sensor_array['cond']['anomaly'] df_anomaly['ph_anom'] = sensor_array['ph']['anomaly'] df_anomaly['do_anom'] = sensor_array['do']['anomaly'] print(df_observed.shape) print(df_raw.shape) print(df_anomaly.shape) ######################################### # LSTM Multivariate Vanilla Model # ######################################### # MODEL CREATION # ######################################### # scales data, reshapes data, builds and trains model, evaluates model results time_steps = 10 samples = 10000 cells = 128 dropout = 0.2 patience = 6 lstm_multivar = modeling_utilities.lstm_multivar(df_observed, df_anomaly, df_raw, time_steps, samples, cells, dropout, patience) # Plot Metrics and Evaluate the Model # plot training loss and validation loss with matplotlib and pyplot plt.plot(lstm_multivar.history.history['loss'], label='Training Loss') plt.plot(lstm_multivar.history.history['val_loss'], label='Validation Loss') plt.legend() plt.show() # DETERMINE THRESHOLD AND DETECT ANOMALIES # ######################################### residuals = pd.DataFrame(lstm_multivar.test_residuals) predictions = pd.DataFrame(lstm_multivar.predictions) residuals.index = df_observed[time_steps:].index predictions.index = df_observed[time_steps:].index window_sz = [40, 40, 40, 40] alpha = [0.0001, 0.0001, 0.0001, 0.001] min_range = [0.25, 5, 0.01, 0.15] threshold = [] for i in range(0, lstm_multivar.test_residuals.shape[1]): threshold_df = anomaly_utilities.set_dynamic_threshold(residuals.iloc[:, i], window_sz[i], alpha[i], min_range[i]) threshold_df.index = residuals.index threshold.append(threshold_df) plt.figure() # plt.plot(df['raw'], 'b', label='original data') plt.plot(residuals.iloc[:, i], 'b', label='residuals') plt.plot(threshold[i]['low'], 'c', label='thresh_low') plt.plot(threshold[i]['high'], 'm', mfc='none', label='thresh_high') plt.legend() plt.ylabel(sensor[i]) plt.show() observed = df_observed[time_steps:] detections_array = [] for i in range(0, observed.shape[1]): detections_df = anomaly_utilities.detect_anomalies(observed.iloc[:, i], lstm_multivar.predictions.iloc[:, i], lstm_multivar.test_residuals.iloc[:, i], threshold[i], summary=True) detections_array.append(detections_df) # Use events function to widen and number anomalous events df_array = [] for i in range(0, len(detections_array)): all_data = [] all_data = sensor_array[sensor[i]].iloc[time_steps:] all_data['labeled_event'] = anomaly_utilities.anomaly_events(all_data['labeled_anomaly']) all_data['detected_anomaly'] = detections_array[i]['anomaly'] all_data['detected_event'] = anomaly_utilities.anomaly_events(all_data['detected_anomaly']) df_array.append(all_data) # DETERMINE METRICS # ######################################### anomaly_utilities.compare_events(df_array[0]) temp_metrics = anomaly_utilities.metrics(df_array[0]) anomaly_utilities.compare_events(df_array[1]) cond_metrics = anomaly_utilities.metrics(df_array[1]) anomaly_utilities.compare_events(df_array[2]) ph_metrics = anomaly_utilities.metrics(df_array[2]) anomaly_utilities.compare_events(df_array[3]) do_metrics = anomaly_utilities.metrics(df_array[3]) # OUTPUT RESULTS # ######################################### print('\n\n\nScript report:\n') print('Sensor: temp') print('Year: ' + str(year)) # print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold)) print('PPV = %f' % temp_metrics.prc) print('NPV = %f' % temp_metrics.npv) print('Acc = %f' % temp_metrics.acc) print('TP = %i' % temp_metrics.true_positives) print('TN = %i' % temp_metrics.true_negatives) print('FP = %i' % temp_metrics.false_positives) print('FN = %i' % temp_metrics.false_negatives) print('F1 = %f' % temp_metrics.f1) print('F2 = %f' % temp_metrics.f2) print('\n\n\nScript report:\n') print('Sensor: cond') print('Year: ' + str(year)) # print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold)) print('PPV = %f' % cond_metrics.prc) print('NPV = %f' % cond_metrics.npv) print('Acc = %f' % cond_metrics.acc) print('TP = %i' % cond_metrics.true_positives) print('TN = %i' % cond_metrics.true_negatives) print('FP = %i' % cond_metrics.false_positives) print('FN = %i' % cond_metrics.false_negatives) print('F1 = %f' % cond_metrics.f1) print('F2 = %f' % cond_metrics.f2) print('\n\n\nScript report:\n') print('Sensor: ph') print('Year: ' + str(year)) # print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold)) print('PPV = %f' % ph_metrics.prc) print('NPV = %f' % ph_metrics.npv) print('Acc = %f' % ph_metrics.acc) print('TP = %i' % ph_metrics.true_positives) print('TN = %i' % ph_metrics.true_negatives) print('FP = %i' % ph_metrics.false_positives) print('FN = %i' % ph_metrics.false_negatives) print('F1 = %f' % ph_metrics.f1) print('F2 = %f' % ph_metrics.f2) print('\n\n\nScript report:\n') print('Sensor: do') print('Year: ' + str(year)) # print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold)) print('PPV = %f' % do_metrics.prc) print('NPV = %f' % do_metrics.npv) print('Acc = %f' % do_metrics.acc) print('TP = %i' % do_metrics.true_positives) print('TN = %i' % do_metrics.true_negatives) print('FP = %i' % do_metrics.false_positives) print('FN = %i' % do_metrics.false_negatives) print('F1 = %f' % do_metrics.f1) print('F2 = %f' % do_metrics.f2) # GENERATE PLOTS # ######################################### for i in range(0, len(sensor)): plt.figure() plt.plot(df_raw[df_raw.columns[i]], 'b', label='original data') #plt.plot(df_observed[df_observed.columns[i]], 'm', label='corrected data' ) plt.plot(detections_array[i]['prediction'], 'c', label='predicted values') plt.plot(sensor_array[sensor[i]]['raw'][sensor_array[sensor[i]]['labeled_anomaly']], 'mo', mfc='none', label='technician labeled anomalies') plt.plot(detections_array[i]['prediction'][detections_array[i]['anomaly']], 'r+', label='machine detected anomalies') plt.legend() plt.ylabel(sensor[i]) plt.show() ######################################### # LSTM Multivariate Bidirectional Model # ######################################### # Model creation # ######################################### # scales data, reshapes data, builds and trains model, evaluates model results time_steps = 10 samples = 10000 cells = 128 dropout = 0.2 patience = 6 lstm_multivar_bidir = modeling_utilities.lstm_multivar_bidir(df_observed, df_anomaly, df_raw, time_steps, samples, cells, dropout, patience) # Plot Metrics and Evaluate the Model # plot training loss and validation loss with matplotlib and pyplot plt.plot(lstm_multivar_bidir.history.history['loss'], label='Training Loss') plt.plot(lstm_multivar_bidir.history.history['val_loss'], label='Validation Loss') plt.legend() plt.show() # DETERMINE THRESHOLD AND DETECT ANOMALIES # ######################################### residuals = pd.DataFrame(lstm_multivar_bidir.test_residuals) residuals.index = df_observed[time_steps:-time_steps].index window_sz = [40, 40, 40, 40] alpha = [0.0001, 0.0001, 0.0001, 0.001] min_range = [0.25, 5, 0.01, 0.15] threshold = [] for i in range(0, lstm_multivar_bidir.test_residuals.shape[1]): threshold_df = anomaly_utilities.set_dynamic_threshold(residuals.iloc[:, i], window_sz[i], alpha[i], min_range[i]) threshold_df.index = residuals.index threshold.append(threshold_df) plt.figure() # plt.plot(df['raw'], 'b', label='original data') plt.plot(residuals.iloc[:, i], 'b', label='residuals') plt.plot(threshold[i]['low'], 'c', label='thresh_low') plt.plot(threshold[i]['high'], 'm', mfc='none', label='thresh_high') plt.legend() plt.ylabel(sensor[i]) plt.show() observed = df_observed[time_steps:-time_steps] detections_array = [] for i in range(0, observed.shape[1]): detections_df = anomaly_utilities.detect_anomalies(observed.iloc[:, i], lstm_multivar_bidir.predictions.iloc[:, i], lstm_multivar_bidir.test_residuals.iloc[:, i], threshold[i], summary=True) detections_array.append(detections_df) # Use events function to widen and number anomalous events df_array = [] for i in range(0, len(detections_array)): all_data = [] all_data = sensor_array[sensor[i]].iloc[time_steps:] all_data['labeled_event'] = anomaly_utilities.anomaly_events(all_data['labeled_anomaly']) all_data['detected_anomaly'] = detections_array[i]['anomaly'] all_data['detected_event'] = anomaly_utilities.anomaly_events(all_data['detected_anomaly']) df_array.append(all_data) # DETERMINE METRICS # ######################################### anomaly_utilities.compare_events(df_array[0]) temp_metrics = anomaly_utilities.metrics(df_array[0]) anomaly_utilities.compare_events(df_array[1]) cond_metrics = anomaly_utilities.metrics(df_array[1]) anomaly_utilities.compare_events(df_array[2]) ph_metrics = anomaly_utilities.metrics(df_array[2]) anomaly_utilities.compare_events(df_array[3]) do_metrics = anomaly_utilities.metrics(df_array[3]) # OUTPUT RESULTS # ######################################### print('\n\n\nScript report:\n') print('Sensor: temp') print('Year: ' + str(year)) # print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold)) print('PPV = %f' % temp_metrics.prc) print('NPV = %f' % temp_metrics.npv) print('Acc = %f' % temp_metrics.acc) print('TP = %i' % temp_metrics.true_positives) print('TN = %i' % temp_metrics.true_negatives) print('FP = %i' % temp_metrics.false_positives) print('FN = %i' % temp_metrics.false_negatives) print('F1 = %f' % temp_metrics.f1) print('F2 = %f' % temp_metrics.f2) print('\n\n\nScript report:\n') print('Sensor: cond') print('Year: ' + str(year)) # print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold)) print('PPV = %f' % cond_metrics.prc) print('NPV = %f' % cond_metrics.npv) print('Acc = %f' % cond_metrics.acc) print('TP = %i' % cond_metrics.true_positives) print('TN = %i' % cond_metrics.true_negatives) print('FP = %i' % cond_metrics.false_positives) print('FN = %i' % cond_metrics.false_negatives) print('F1 = %f' % cond_metrics.f1) print('F2 = %f' % cond_metrics.f2) print('\n\n\nScript report:\n') print('Sensor: ph') print('Year: ' + str(year)) # print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold)) print('PPV = %f' % ph_metrics.prc) print('NPV = %f' % ph_metrics.npv) print('Acc = %f' % ph_metrics.acc) print('TP = %i' % ph_metrics.true_positives) print('TN = %i' % ph_metrics.true_negatives) print('FP = %i' % ph_metrics.false_positives) print('FN = %i' % ph_metrics.false_negatives) print('F1 = %f' % ph_metrics.f1) print('F2 = %f' % ph_metrics.f2) print('\n\n\nScript report:\n') print('Sensor: do') print('Year: ' + str(year)) # print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold)) print('PPV = %f' % do_metrics.prc) print('NPV = %f' % do_metrics.npv) print('Acc = %f' % do_metrics.acc) print('TP = %i' % do_metrics.true_positives) print('TN = %i' % do_metrics.true_negatives) print('FP = %i' % do_metrics.false_positives) print('FN = %i' % do_metrics.false_negatives) print('F1 = %f' % do_metrics.f1) print('F2 = %f' % do_metrics.f2) # GENERATE PLOTS # ######################################### for i in range(0, len(sensor)): plt.figure() plt.plot(df_raw[df_raw.columns[i]], 'b', label='original data') # plt.plot(df_observed[df_observed.columns[i]], 'm', label='corrected data' ) plt.plot(detections_array[i]['prediction'], 'c', label='predicted values') plt.plot(sensor_array[sensor[i]]['raw'][sensor_array[sensor[i]]['labeled_anomaly']], 'mo', mfc='none', label='technician labeled anomalies') plt.plot(detections_array[i]['prediction'][detections_array[i]['anomaly']], 'r+', label='machine detected anomalies') plt.legend() plt.ylabel(sensor[i]) plt.show() print("\n LSTM script end.")
38.907407
194
0.671857
2,034
14,707
4.66765
0.117502
0.043817
0.025068
0.024647
0.821045
0.811671
0.793448
0.78102
0.78102
0.78102
0
0.015823
0.114775
14,707
377
195
39.01061
0.713419
0.160672
0
0.740458
0
0
0.156203
0
0
0
0
0
0
1
0
false
0
0.038168
0
0.038168
0.389313
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
357c732ea7136d2889cda027fe0436f8645a4363
1,977
py
Python
src/viola/testing.py
mmiki21/Viola-SV
10fe2c326749024551d3e41f900b37e20dbde35c
[ "Apache-2.0" ]
13
2021-05-03T18:43:25.000Z
2022-03-26T04:57:01.000Z
src/viola/testing.py
mmiki21/Viola-SV
10fe2c326749024551d3e41f900b37e20dbde35c
[ "Apache-2.0" ]
6
2021-04-02T08:04:23.000Z
2022-03-24T11:34:43.000Z
src/viola/testing.py
mmiki21/Viola-SV
10fe2c326749024551d3e41f900b37e20dbde35c
[ "Apache-2.0" ]
3
2022-02-23T03:36:12.000Z
2022-03-07T07:05:00.000Z
import viola from viola.core.bedpe import Bedpe from viola.core.vcf import Vcf import pandas as pd def assert_bedpe_equal(left, right, check_like=True): ls_left_table = left.table_list ls_right_table = right.table_list assert len(ls_left_table) == len(ls_right_table) assert set(ls_left_table) == set(ls_right_table) for tablename in ls_left_table: df_left = left.get_table(tablename) df_right = right.get_table(tablename) id_left = df_left.columns[0] id_right = df_right.columns[0] df_left.sort_values(id_left, inplace=True, ignore_index=True) df_right.sort_values(id_right, inplace=True, ignore_index=True) if df_left.empty & df_right.empty: continue try: pd.testing.assert_frame_equal(df_left, df_right, check_like=check_like, check_exact=True) except AssertionError: print('\nwhen asserting {} table, following error occured!'.format(tablename)) raise def assert_vcf_equal(left, right, check_like=True): ls_left_table = left.table_list ls_right_table = right.table_list assert len(ls_left_table) == len(ls_right_table) assert set(ls_left_table) == set(ls_right_table) for tablename in ls_left_table: df_left = left.get_table(tablename) df_right = right.get_table(tablename) id_left = df_left.columns[0] id_right = df_right.columns[0] df_left.sort_values(id_left, inplace=True, ignore_index=True) df_right.sort_values(id_right, inplace=True, ignore_index=True) if df_left.empty & df_right.empty: continue try: pd.testing.assert_frame_equal(df_left, df_right, check_like=check_like, check_exact=True) except AssertionError: print('\nwhen asserting {} table, following error occured!'.format(tablename)) raise # assert metadata assert left._metadata == right._metadata
39.54
101
0.687911
281
1,977
4.519573
0.188612
0.070866
0.069291
0.069291
0.880315
0.880315
0.880315
0.880315
0.880315
0.880315
0
0.002616
0.226606
1,977
49
102
40.346939
0.827992
0.007587
0
0.837209
0
0
0.052041
0
0
0
0
0
0.302326
1
0.046512
false
0
0.093023
0
0.139535
0.046512
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
3595b042857ebc343975566768abd9ecefaa9c77
3,223
py
Python
tests/test_deque.py
EndreoT/DataStructures
3359c142a3a3b1eb2e275a90544973dd709b0d29
[ "MIT" ]
null
null
null
tests/test_deque.py
EndreoT/DataStructures
3359c142a3a3b1eb2e275a90544973dd709b0d29
[ "MIT" ]
null
null
null
tests/test_deque.py
EndreoT/DataStructures
3359c142a3a3b1eb2e275a90544973dd709b0d29
[ "MIT" ]
null
null
null
import unittest from deque import Deque from node import Node class TestQueue(unittest.TestCase): def setUp(self): self.q = Deque() def test_empty_deque(self): self.assertEqual(self.q.get_front(), None) self.assertEqual(self.q.get_rear(), None) self.assertEqual(self.q.get_front(), self.q.get_rear()) self.assertEqual(self.q.get_size(), 0) self.assertEqual(self.q.remove_front(), None) def test_one_element_deque(self): q = Deque(1) self.assertEqual(q.get_front().get_value(), 1) self.assertEqual(q.get_rear().get_value(), 1) self.assertEqual(type(q.get_front()), Node) self.assertEqual(type(q.get_rear()), Node) self.assertEqual(q.get_rear(), q.get_front()) self.assertEqual(q.get_rear().get_previous(), None) self.assertEqual(q.get_rear().get_next(), None) self.assertEqual(q.get_front().get_next(), None) self.assertEqual(q.get_front().get_previous(), None) self.assertEqual(q.get_size(), 1) def test_add_front(self): for i in range(1, 4): self.q.add_front(i) self.assertEqual(self.q.get_front().get_value(), 3) self.assertEqual(self.q.get_rear().get_next().get_value(), 2) self.assertEqual(self.q.get_front().get_next(), None) self.assertEqual(self.q.get_front().get_previous().get_value(), 2) self.assertEqual(self.q.get_size(), 3) def test_add_rear(self): for i in range(1, 4): self.q.add_rear(i) self.assertEqual(self.q.get_front().get_value(), 1) self.assertEqual(self.q.get_rear().get_next().get_value(), 2) self.assertEqual(self.q.get_front().get_next(), None) self.assertEqual(self.q.get_front().get_previous().get_value(), 2) self.assertEqual(self.q.get_size(), 3) def test_remove_front(self): for i in range(1, 8): self.q.add_front(i) removed = self.q.remove_front() self.assertEqual(self.q.get_size(), 6) self.assertEqual(self.q.get_front().get_value(), 6) self.assertEqual(self.q.get_front().get_next(), None) self.assertEqual(self.q.get_front().get_previous().get_value(), 5) self.assertEqual(removed.get_value(), 7) self.assertEqual(removed.get_previous(), None) def test_remove_rear(self): for i in range(1, 8): self.q.add_rear(i) removed = self.q.remove_rear() self.assertEqual(self.q.get_size(), 6) self.assertEqual(self.q.get_rear().get_value(), 6) self.assertEqual(self.q.get_rear().get_previous(), None) self.assertEqual(self.q.get_rear().get_next().get_value(), 5) self.assertEqual(removed.get_value(), 7) self.assertEqual(removed.get_previous(), None) def test_add_both(self): self.q.add_front(2) self.q.add_rear(1) self.assertEqual(self.q.get_rear().get_value(), 1) self.assertEqual(self.q.get_front().get_value(), 2) self.assertEqual(self.q.get_rear().get_next().get_value(), 2) self.assertEqual(self.q.get_front().get_previous().get_value(), 1) if __name__ == "__main__": unittest.main()
37.917647
74
0.635743
472
3,223
4.116525
0.097458
0.316521
0.264025
0.277921
0.833762
0.776634
0.744725
0.66701
0.539887
0.464231
0
0.013656
0.204778
3,223
84
75
38.369048
0.74444
0
0
0.347826
0
0
0.002482
0
0
0
0
0
0.594203
1
0.115942
false
0
0.043478
0
0.173913
0
0
0
0
null
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
1
0
0
0
0
0
0
0
0
0
7
57a56459a1620898f667910fbc474dac679355b2
23,983
py
Python
pycqed/measurement/waveform_control_CC/QWG_fluxing_seqs.py
nuttamas/PycQED_py3
1ee35c7428d36ed42ba4afb5d4bda98140b2283e
[ "MIT" ]
60
2016-08-03T10:00:18.000Z
2021-11-10T11:46:16.000Z
pycqed/measurement/waveform_control_CC/QWG_fluxing_seqs.py
nuttamas/PycQED_py3
1ee35c7428d36ed42ba4afb5d4bda98140b2283e
[ "MIT" ]
512
2016-08-03T17:10:02.000Z
2022-03-31T14:03:43.000Z
pycqed/measurement/waveform_control_CC/QWG_fluxing_seqs.py
nuttamas/PycQED_py3
1ee35c7428d36ed42ba4afb5d4bda98140b2283e
[ "MIT" ]
34
2016-10-19T12:00:52.000Z
2022-03-19T04:43:26.000Z
from pycqed.utilities.general import mopen from os.path import join, dirname import numpy as np base_qasm_path = join(dirname(__file__), 'qasm_files') from pycqed.measurement.waveform_control_CC.multi_qubit_qasm_seqs \ import cal_points_2Q def ramZ_flux_latency(q0_name, wait_after_flux=20): """ Sequence designed to calibrate the delay between the QWG_trigger and the start of the flux pulse Consists of a single point. Intended to change the latency parameter in the configuration that is used in compilation. """ filename = join(base_qasm_path, 'RamZ_latency_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \n'.format(q0_name)) # simultaneous MW and flux pulse qasm_file.writelines('\ninit_all\n') qasm_file.writelines('X90 {} \n'.format(q0_name)) qasm_file.writelines('square {}\n'.format(q0_name)) qasm_file.writelines('I {}\n'.format(wait_after_flux)) qasm_file.writelines('X90 {}\n'.format(q0_name)) qasm_file.writelines('RO {} \n'.format(q0_name)) qasm_file.close() return qasm_file def chevron_block_seq(q0_name, q1_name, no_of_points, excite_q1=False, wait_after_trigger=40e-9, wait_during_flux=400e-9, clock_cycle=1e-9, RO_target='all', mw_pulse_duration=40e-9, cal_points=True): ''' N.B. this sequence has been edited for compatibility with the XFU compiler Sequence for measuring a block of a chevron, i.e. using different codewords for different pulse lengths. Args: q0, q1 (str): names of the addressed qubits. q0 is the pulse that experiences the flux pulse. RO_target (str): can be q0, q1, or 'all' excite_q1 (bool): choose whether to excite q1, thus choosing between the |01> <-> |10> and the |11> <-> |20> swap. wait_after_trigger (float): delay time in seconds after sending the trigger for the flux pulse clock_cycle (float): period of the internal AWG clock wait_time (int): wait time between triggering QWG and RO cal_points (bool): whether to use calibration points or not ''' filename = join(base_qasm_path, 'chevron_block_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0_name, q1_name)) for i in range(no_of_points): qasm_file.writelines('\ninit_all\n') qasm_file.writelines('QWG_trigger_{} {}\n'.format(i, q0_name)) if excite_q1: wait_after_trigger -= mw_pulse_duration qasm_file.writelines('X180 {}\n'.format(q0_name)) if excite_q1: qasm_file.writelines('X180 {}\n'.format(q1_name)) qasm_file.writelines('CZ {} {}\n'.format(q0_name, q1_name)) if excite_q1: # q0 is rotated to ground-state to have better contrast # (|0> and |2> instead of |1> and |2>) qasm_file.writelines('X180 {}\n'.format(q0_name)) if RO_target == 'all': qasm_file.writelines('RO {} | RO {} \n'.format(q0_name, q1_name)) else: qasm_file.writelines('RO {} \n'.format(RO_target)) if cal_points: # Add calibration pulses cal_pulses = [] for seq in cal_points_2Q: cal_pulses += [[seq[0], seq[1], 'RO ' + RO_target + '\n']] qasm_file.close() return qasm_file def SWAPN(q0_name, q1_name, nr_pulses: list, excite_q1=False, RO_target='all', cal_points=True): ''' Args: q0, q1 (str): names of the addressed qubits. q0 is the pulse that experiences the flux pulse. RO_target (str): can be q0, q1, or 'all' excite_q1 (bool): choose whether to excite q1, thus choosing between the |01> <-> |10> and the |11> <-> |20> swap. cal_points (bool): whether to use calibration points or not ''' filename = join(base_qasm_path, 'chevron_block_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0_name, q1_name)) for i, N in enumerate(nr_pulses): qasm_file.writelines('\ninit_all\n') qasm_file.writelines('QWG_trigger_{} {}\n'.format(i, q0_name)) qasm_file.writelines('X180 {}\n'.format(q0_name)) if excite_q1: qasm_file.writelines('X180 {}\n'.format(q1_name)) for n in range(N): qasm_file.writelines('square {} \n'.format(q0_name)) if excite_q1: # q0 is rotated to ground-state to have better contrast # (|0> and |2> instead of |1> and |2>) qasm_file.writelines('X180 {}\n'.format(q0_name)) qasm_file.writelines('RO {} \n'.format(RO_target)) if cal_points: # Add calibration pulses cal_pulses = [] for seq in cal_points_2Q: cal_pulses += [[seq[0].format(q0_name) + seq[1].format(q1_name) + 'RO {} \n'.format(RO_target)]] for seq in cal_pulses: qasm_file.writelines('\ninit_all\n') for p in seq: qasm_file.writelines(p) qasm_file.close() return qasm_file def CZ_calibration_seq(q0, q1, RO_target='all', vary_single_q_phase=True, cases=('no_excitation', 'excitation')): ''' Sequence used to calibrate flux pulses for CZ gates. Timing of the sequence: q0: X90 C-Phase Rphi90 RO q1: (X180) -- (X180) RO Args: q0, q1 (str): names of the addressed qubits RO_target (str): can be q0, q1, or 'all' excitations (bool/str): can be True, False, or 'both_cases' ''' filename = join(base_qasm_path, 'CZ_calibration_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0, q1)) for case in cases: qasm_file.writelines('\ninit_all\n') if case == 'excitation': qasm_file.writelines('X180 {} | '.format(q1)) qasm_file.writelines('X90 {}\n'.format(q0)) # temporary workaround to deal with limitation in the QASM config # qasm_file.writelines('CZ {} \n'.format(q0)) qasm_file.writelines('CZ {} {}\n'.format(q0, q1)) if case == 'excitation': qasm_file.writelines('X180 {} | '.format(q1)) if vary_single_q_phase: qasm_file.writelines('Rphi90 {}\n'.format(q0)) else: qasm_file.writelines('mX90 {}\n'.format(q0)) if 'RO_target' == 'all': qasm_file.writelines('RO {} | RO {} \n'.format(q0, q1)) else: qasm_file.writelines('RO {} \n'.format(RO_target)) qasm_file.close() return qasm_file def two_qubit_tomo_bell(bell_state, q0, q1, RO_target='all'): ''' Two qubit bell state tomography. Args: bell_state (int): index of prepared bell state 0 : |00>-|11> 1 : |00>+|11> 2 : |01>-|10> 3 : |01>+|10> q0, q1 (str): names of the target qubits RO_target (str): can be q0, q1, or 'all' ''' if RO_target == 'all': # This is a bit of a hack as RO all qubits is the same instruction # as any specific qubit RO_target = q0 tomo_pulses = ['I ', 'X180 ', 'Y90 ', 'mY90 ', 'X90 ', 'mX90 '] tomo_list_q0 = [] tomo_list_q1 = [] for tp in tomo_pulses: tomo_list_q0 += [tp + q0] tomo_list_q1 += [tp + q1] tomo_list_q0[0] = 'I {}'.format(q0) tomo_list_q1[0] = 'I {}'.format(q1) # Choose a bell state and set the corresponding preparation pulses if bell_state % 10 == 0: # |Phi_m>=|00>-|11> prep_pulse_q0 = 'Y90 {}'.format(q0) prep_pulse_q1 = 'Y90 {}'.format(q1) elif bell_state % 10 == 1: # |Phi_p>=|00>+|11> prep_pulse_q0 = 'mY90 {}'.format(q0) prep_pulse_q1 = 'Y90 {}'.format(q1) elif bell_state % 10 == 2: # |Psi_m>=|01>-|10> prep_pulse_q0 = 'Y90 {}'.format(q0) prep_pulse_q1 = 'mY90 {}'.format(q1) elif bell_state % 10 == 3: # |Psi_p>=|01>+|10> prep_pulse_q0 = 'mY90 {}'.format(q0) prep_pulse_q1 = 'mY90 {}'.format(q1) else: raise ValueError('Bell state {} is not defined.'.format(bell_state)) after_pulse = 'mY90 {}\n'.format(q1) # Disable preparation pulse on one or the other qubit for debugging if bell_state//10 == 1: prep_pulse_q1 = 'I {}'.format(q0) elif bell_state//10 == 2: prep_pulse_q0 = 'I {}'.format(q1) # Write tomo sequence filename = join(base_qasm_path, 'two_qubit_tomo_bell.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0, q1)) for p_q1 in tomo_list_q1: for p_q0 in tomo_list_q0: qasm_file.writelines('\ninit_all\n') qasm_file.writelines('{} | {} \n'.format(prep_pulse_q0, prep_pulse_q1)) qasm_file.writelines('CZ {} {} \n'.format(q0, q1)) qasm_file.writelines(after_pulse) qasm_file.writelines('{} | {}\n'.format(p_q1, p_q0)) qasm_file.writelines('RO ' + RO_target + ' \n') # Add calibration pulses cal_pulses = [] # every calibration point is repeated 7 times. This is copied from the # script for Tektronix driven qubits. I do not know if this repetition # is important or even necessary here. for seq in cal_points_2Q: cal_pulses += [[seq[0].format(q0), seq[1].format(q1), 'RO ' + RO_target + '\n']] * 7 for seq in cal_pulses: qasm_file.writelines('\ninit_all\n') for p in seq: qasm_file.writelines(p) qasm_file.close() return qasm_file def grover_seq(q0_name, q1_name, RO_target='all', precompiled_flux=True, cal_points: bool=True): ''' Writes the QASM sequence for Grover's algorithm on two qubits. Sequence: q0: G0 - - mY90 - - mY90 - RO CZ_ij CZ q1: G1 - - mY90 - - mY90 - RO whit all combinations of (ij) = omega. G0 and G1 are Y90 or Y90, depending on the (ij). Args: q0_name, q1_name (string): Names of the qubits to which the sequence is applied. RO_target (string): Readout target. Can be a qubit name or 'all'. precompiled_flux (bool): Determies if the full waveform for the flux pulses is precompiled, thus only needing one trigger at the start, or if every flux pulse should be triggered individually. cal_points (bool): Whether to add calibration points. Returns: qasm_file: a reference to the new QASM file object. ''' if not precompiled_flux: raise NotImplementedError('Currently only precompiled flux pulses ' 'are supported.') filename = join(base_qasm_path, 'Grover_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \n'.format(q0_name)) qasm_file.writelines('qubit {} \n'.format(q1_name)) if RO_target == 'all': RO_line = 'RO {} | RO {}\n'.format(q0_name, q1_name) else: RO_line = 'RO {} \n'.format(RO_target) for G0 in ['Y90', 'mY90']: for G1 in ['Y90', 'mY90']: qasm_file.writelines('\ninit_all\n') qasm_file.writelines('{} {} | {} {}\n'.format(G0, q0_name, G1, q1_name)) qasm_file.writelines('grover_CZ {} {}\n'.format(q0_name, q1_name)) qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name, q1_name)) qasm_file.writelines('cz {} {}\n'.format(q0_name, q1_name)) qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name, q1_name)) qasm_file.writelines(RO_line) # Add calibration points if cal_points: cal_pulses = [] for seq in cal_points_2Q: cal_pulses += [[seq[0].format(q0_name), seq[1].format(q1_name), RO_line]] for seq in cal_pulses: qasm_file.writelines('\ninit_all\n') for p in seq: qasm_file.writelines(p) qasm_file.close() return qasm_file def grover_tomo_seq(q0_name, q1_name, omega, RO_target='all', precompiled_flux=True): ''' Writes the QASM sequence to take a state tomography of the output state of Grover's algorithm on two qubits. Sequence: q0: G0 - - mY90 - - mY90 - RO CZ_ij CZ q1: G1 - - mY90 - - mY90 - RO where (ij) is the binary representation of omega. G0 and G1 are Y90 or Y90, depending on the (ij). Args: q0_name, q1_name (string): Names of the qubits to which the sequence is applied. omega (int): Deterines which (ij) for the CZ_ij. RO_target (string): Readout target. Can be a qubit name or 'all'. precompiled_flux (bool): Determies if the full waveform for the flux pulses is precompiled, thus only needing one trigger at the start, or if every flux pulse should be triggered individually. Returns: qasm_file: a reference to the new QASM file object. ''' if not precompiled_flux: raise NotImplementedError('Currently only precompiled flux pulses ' 'are supported.') tomo_pulses = ['I ', 'X180 ', 'Y90 ', 'mY90 ', 'X90 ', 'mX90 '] tomo_list_q0 = [] tomo_list_q1 = [] for tp in tomo_pulses: tomo_list_q0 += [tp + q0_name] tomo_list_q1 += [tp + q1_name] if omega == 0: G0 = 'Y90' G1 = 'Y90' elif omega == 1: G0 = 'Y90' G1 = 'mY90' elif omega == 2: G0 = 'mY90' G1 = 'Y90' elif omega == 3: G0 = 'mY90' G1 = 'mY90' else: raise ValueError('omega must be in [0, 3]') if RO_target == 'all': RO_line = 'RO {} | RO {}\n'.format(q0_name, q1_name) else: RO_line = 'RO {} \n'.format(RO_target) filename = join(base_qasm_path, 'Grover_tomo_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \n'.format(q0_name)) qasm_file.writelines('qubit {} \n'.format(q1_name)) for p_q1 in tomo_list_q1: for p_q0 in tomo_list_q0: qasm_file.writelines('\ninit_all\n') qasm_file.writelines('{} {} | {} {}\n'.format(G0, q0_name, G1, q1_name)) qasm_file.writelines('grover_CZ {} {}\n'.format(q0_name, q1_name)) qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name, q1_name)) qasm_file.writelines('cz {} {}\n'.format(q0_name, q1_name)) qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name, q1_name)) qasm_file.writelines('{} | {}\n'.format(p_q1, p_q0)) qasm_file.writelines(RO_line) # Add calibration pulses cal_pulses = [] # every calibration point is repeated 7 times. This is copied from the # script for Tektronix driven qubits. I do not know if this repetition # is important or even necessary here. for seq in cal_points_2Q: cal_pulses += [[seq[0].format(q0_name), seq[1].format(q1_name), RO_line]] * 7 for seq in cal_pulses: qasm_file.writelines('\ninit_all\n') for p in seq: qasm_file.writelines(p) qasm_file.close() return qasm_file def grover_test_seq(q0_name, q1_name, RO_target='all', precompiled_flux=True, cal_points: bool=True): ''' Writes the QASM sequence for Grover's algorithm on two qubits. Sequence: q0: G0 - - mY90 - - mY90 - RO CZ_ij CZ q1: G1 - - mY90 - - mY90 - RO whit all combinations of (ij) = omega. G0 and G1 are Y90 or Y90, depending on the (ij). Args: q0_name, q1_name (string): Names of the qubits to which the sequence is applied. RO_target (string): Readout target. Can be a qubit name or 'all'. precompiled_flux (bool): Determies if the full waveform for the flux pulses is precompiled, thus only needing one trigger at the start, or if every flux pulse should be triggered individually. cal_points (bool): Whether to add calibration points. Returns: qasm_file: a reference to the new QASM file object. ''' if not precompiled_flux: raise NotImplementedError('Currently only precompiled flux pulses ' 'are supported.') filename = join(base_qasm_path, 'Grover_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \n'.format(q0_name)) qasm_file.writelines('qubit {} \n'.format(q1_name)) if RO_target == 'all': RO_line = 'RO {} | RO {}\n'.format(q0_name, q1_name) else: RO_line = 'RO {} \n'.format(RO_target) for G0 in ['Y90', 'mY90']: for G1 in ['Y90', 'mY90']: qasm_file.writelines('\ninit_all\n') qasm_file.writelines('{} {} | {} {}\n'.format(G0, q0_name, G1, q1_name)) qasm_file.writelines('grover_CZ {} {}\n'.format(q0_name, q1_name)) qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name, q1_name)) # qasm_file.writelines('cz {} {}\n'.format(q0_name, q1_name)) # qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name, # q1_name)) qasm_file.writelines(RO_line) # Add calibration points if cal_points: cal_pulses = [] for seq in cal_points_2Q: cal_pulses += [[seq[0].format(q0_name), seq[1].format(q1_name), RO_line]] for seq in cal_pulses: qasm_file.writelines('\ninit_all\n') for p in seq: qasm_file.writelines(p) qasm_file.close() return qasm_file def grover_test_tomo_seq(q0_name, q1_name, omega, RO_target='all', precompiled_flux=True): ''' Test sequence to debug Grover's algorithm. ''' if not precompiled_flux: raise NotImplementedError('Currently only precompiled flux pulses ' 'are supported.') tomo_pulses = ['I ', 'X180 ', 'Y90 ', 'mY90 ', 'X90 ', 'mX90 '] tomo_list_q0 = [] tomo_list_q1 = [] for tp in tomo_pulses: tomo_list_q0 += [tp + q0_name] tomo_list_q1 += [tp + q1_name] if omega == 0: G0 = 'Y90' G1 = 'Y90' elif omega == 1: G0 = 'Y90' G1 = 'mY90' elif omega == 2: G0 = 'mY90' G1 = 'Y90' elif omega == 3: G0 = 'mY90' G1 = 'mY90' else: raise ValueError('omega must be in [0, 3]') if RO_target == 'all': RO_line = 'RO {} | RO {}\n'.format(q0_name, q1_name) else: RO_line = 'RO {} \n'.format(RO_target) filename = join(base_qasm_path, 'Grover_tomo_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \n'.format(q0_name)) qasm_file.writelines('qubit {} \n'.format(q1_name)) for p_q1 in tomo_list_q1: for p_q0 in tomo_list_q0: qasm_file.writelines('\ninit_all\n') qasm_file.writelines('{} {} | {} {}\n'.format(G0, q0_name, G1, q1_name)) qasm_file.writelines('grover_cz {} {}\n'.format(q0_name, q1_name)) qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name, q1_name)) qasm_file.writelines('cz {} {}\n'.format(q0_name, q1_name)) # qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name, # q1_name)) qasm_file.writelines('{} | {}\n'.format(p_q1, p_q0)) qasm_file.writelines(RO_line) # Add calibration pulses cal_pulses = [] # every calibration point is repeated 7 times. This is copied from the # script for Tektronix driven qubits. I do not know if this repetition # is important or even necessary here. for seq in cal_points_2Q: cal_pulses += [[seq[0].format(q0_name), seq[1].format(q1_name), RO_line]] * 7 for seq in cal_pulses: qasm_file.writelines('\ninit_all\n') for p in seq: qasm_file.writelines(p) qasm_file.close() return qasm_file def purity_CZ_seq(q0, q1, RO_target='all'): """ Creates the |00> + |11> Bell state and does a partial tomography in order to determine the purity of both qubits. """ filename = join(base_qasm_path, 'purity_CZ_seq.qasm') qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0, q1)) tomo_list = ['mX90', 'mY90', 'I'] for p_pulse in tomo_list: # Create a Bell state: |00> + |11> qasm_file.writelines('\ninit_all\n') qasm_file.writelines('mY90 {} | Y90 {} \n'.format(q0, q1)) qasm_file.writelines('CZ {} {} \n'.format(q0, q1)) qasm_file.writelines('mY90 {}\n'.format(q1)) # Perform pulses to measure the purity of both qubits qasm_file.writelines('{} {} | {} {}\n'.format(p_pulse, q0, p_pulse, q1)) if RO_target == 'all': qasm_file.writelines('RO {} | RO {} \n'.format(q0, q1)) else: qasm_file.writelines('RO {} \n'.format(RO_target)) qasm_file.close() return qasm_file def purity_N_CZ_seq(q0: str, q1: str, N: int, RO_target: str='all'): """ Creates the |00> + |11> Bell state and does a partial tomography in order to determine the purity of both qubits. """ filename = join(base_qasm_path, 'purity_{}_CZ_seq.qasm'.format(N)) qasm_file = mopen(filename, mode='w') qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0, q1)) tomo_list = ['mX90', 'mY90', 'I'] for i, p_pulse in enumerate(tomo_list): # Create a Bell state: |00> + |11> qasm_file.writelines('\ninit_all\n') qasm_file.writelines('qwg_trigger_{} {}\n'.format(0, q0)) qasm_file.writelines('mY90 {} | Y90 {} \n'.format(q0, q1)) for n in range(N): qasm_file.writelines('dummy_CZ {} {} \n'.format(q0, q1)) qasm_file.writelines('mY90 {}\n'.format(q1)) # Perform pulses to measure the purity of both qubits qasm_file.writelines('{} {} | {} {}\n'.format(p_pulse, q0, p_pulse, q1)) if RO_target == 'all': qasm_file.writelines('RO {} | RO {} \n'.format(q0, q1)) else: qasm_file.writelines('RO {} \n'.format(RO_target)) qasm_file.close() return qasm_file
37.356698
79
0.556686
3,139
23,983
4.052246
0.090156
0.093082
0.154245
0.038836
0.824135
0.809827
0.800943
0.789701
0.777044
0.752594
0
0.041887
0.322103
23,983
641
80
37.414977
0.740497
0.279365
0
0.805405
0
0
0.132134
0.007974
0
0
0
0
0
1
0.02973
false
0
0.010811
0
0.07027
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
57e59a41b444f8b1198e53a31beb8895923cd651
1,175
py
Python
app/map/forms/__init__.py
jg-725/IS219-FlaskAppProject
316aa298eda1bcda766ed085bb6f26ca7da7dfee
[ "BSD-3-Clause" ]
null
null
null
app/map/forms/__init__.py
jg-725/IS219-FlaskAppProject
316aa298eda1bcda766ed085bb6f26ca7da7dfee
[ "BSD-3-Clause" ]
null
null
null
app/map/forms/__init__.py
jg-725/IS219-FlaskAppProject
316aa298eda1bcda766ed085bb6f26ca7da7dfee
[ "BSD-3-Clause" ]
null
null
null
from flask_wtf import FlaskForm from wtforms import validators from wtforms.fields import * class csv_upload(FlaskForm): file = FileField() submit = SubmitField() class create_location_form(FlaskForm): title = StringField('City Name', description="Add The City Name") latitude = StringField('Latitude:', [ validators.DataRequired(), ], description="Positive Value") longitude = StringField('Longitude:', [ validators.DataRequired(), ], description="Negative Value") population = IntegerField('Population:', [ validators.DataRequired(), ], description="Add Population Value ") submit = SubmitField() class edit_location_form(FlaskForm): title = StringField('City Name', description="Edit The City Name") latitude = StringField('Latitude:', [ validators.DataRequired(), ], description=" Positive Value ") longitude = StringField('Longitude:', [ validators.DataRequired(), ], description="Negative Value ") population = IntegerField('Population:', [ validators.DataRequired(), ], description="Add Population Value ") submit = SubmitField()
23.979592
70
0.674894
103
1,175
7.640777
0.330097
0.167726
0.251588
0.066074
0.800508
0.800508
0.800508
0.800508
0.658196
0.658196
0
0
0.203404
1,175
49
71
23.979592
0.840812
0
0
0.7
0
0
0.181973
0
0
0
0
0
0
1
0
false
0
0.1
0
0.6
0
0
0
0
null
0
1
0
1
1
1
1
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
1
0
0
7
17b57f56133daabf92fa3b1bc58aa9bfbe85e35a
4,137
py
Python
chsdi/models/vector/kogis.py
procrastinatio/mf-chsdi3
206e9362860376338cc83ded902790dc216e2563
[ "BSD-3-Clause" ]
null
null
null
chsdi/models/vector/kogis.py
procrastinatio/mf-chsdi3
206e9362860376338cc83ded902790dc216e2563
[ "BSD-3-Clause" ]
null
null
null
chsdi/models/vector/kogis.py
procrastinatio/mf-chsdi3
206e9362860376338cc83ded902790dc216e2563
[ "BSD-3-Clause" ]
null
null
null
# -*- coding: utf-8 -*- from sqlalchemy import Column, Integer from sqlalchemy.types import Numeric, Unicode from chsdi.models import register, bases from chsdi.models.vector import Vector, Geometry2D Base = bases['kogis'] class Agnes(Base, Vector): __tablename__ = 'agnes' __table_args__ = ({'schema': 'fpds', 'autoload': False}) __template__ = 'templates/htmlpopup/agnes.mako' __bodId__ = 'ch.swisstopo.fixpunkte-agnes' __label__ = 'id' id = Column('no', Unicode, primary_key=True) url = Column('url', Unicode) bgdi_id = Column('bgdi_id', Integer) the_geom = Column(Geometry2D) register('ch.swisstopo.fixpunkte-agnes', Agnes) class FixpunkteLfp1(Base, Vector): __tablename__ = 'punkt_lage_lfp1' __table_args__ = ({'schema': 'fpds', 'autoload': False}) __template__ = 'templates/htmlpopup/fixpunkte.mako' __bodId__ = 'ch.swisstopo.fixpunkte-lfp1' __label__ = 'id' id = Column('pointid', Unicode, primary_key=True) punktname = Column('punktname', Unicode) nummer = Column('nummer', Unicode) status = Column('status', Unicode) nbident = Column('nbident', Unicode) x03 = Column('x03', Numeric) y03 = Column('y03', Numeric) n95 = Column('n95', Numeric) e95 = Column('e95', Numeric) h02 = Column('h02', Numeric) zugang = Column('zugang', Unicode) url = Column('url', Unicode) bgdi_created = Column('bgdi_created', Unicode) the_geom = Column(Geometry2D) register('ch.swisstopo.fixpunkte-lfp1', FixpunkteLfp1) class FixpunkteLfp2(Base, Vector): __tablename__ = 'punkt_lage_lfp2' __table_args__ = ({'schema': 'fpds', 'autoload': False}) __template__ = 'templates/htmlpopup/fixpunkte.mako' __bodId__ = 'ch.swisstopo.fixpunkte-lfp2' __label__ = 'id' id = Column('pointid', Unicode, primary_key=True) nbident = Column('nbident', Unicode) punktname = Column('punktname', Unicode) status = Column('status', Unicode) nummer = Column('nummer', Unicode) x03 = Column('x03', Numeric) y03 = Column('y03', Numeric) n95 = Column('n95', Numeric) e95 = Column('e95', Numeric) h02 = Column('h02', Numeric) zugang = Column('zugang', Unicode) url = Column('url', Unicode) bgdi_created = Column('bgdi_created', Unicode) the_geom = Column(Geometry2D) register('ch.swisstopo.fixpunkte-lfp2', FixpunkteLfp2) class FixpunkteHfp1(Base, Vector): __tablename__ = 'punkt_hoehe_hfp1' __table_args__ = ({'schema': 'fpds', 'autoload': True}) __template__ = 'templates/htmlpopup/fixpunkte.mako' __bodId__ = 'ch.swisstopo.fixpunkte-hfp1' __label__ = 'id' id = Column('pointid', Unicode, primary_key=True) bgdi_label = Column('bgdi_label', Unicode) nbident = Column('nbident', Unicode) punktname = Column('punktname', Unicode) status = Column('status', Unicode) nummer = Column('nummer', Unicode) x03 = Column('x03', Numeric) y03 = Column('y03', Numeric) n95 = Column('n95', Numeric) e95 = Column('e95', Numeric) h02 = Column('h02', Numeric) zugang = Column('zugang', Unicode) url = Column('url', Unicode) bgdi_created = Column('bgdi_created', Unicode) the_geom = Column(Geometry2D) register('ch.swisstopo.fixpunkte-hfp1', FixpunkteHfp1) class FixpunkteHfp2(Base, Vector): __tablename__ = 'punkt_hoehe_hfp2' __table_args__ = ({'schema': 'fpds', 'autoload': True}) __template__ = 'templates/htmlpopup/fixpunkte.mako' __bodId__ = 'ch.swisstopo.fixpunkte-hfp2' __label__ = 'id' id = Column('pointid', Unicode, primary_key=True) nbident = Column('nbident', Unicode) punktname = Column('punktname', Unicode) status = Column('status', Unicode) nummer = Column('nummer', Unicode) x03 = Column('x03', Numeric) y03 = Column('y03', Numeric) n95 = Column('n95', Numeric) e95 = Column('e95', Numeric) h02 = Column('h02', Numeric) zugang = Column('zugang', Unicode) url = Column('url', Unicode) bgdi_created = Column('bgdi_created', Unicode) the_geom = Column(Geometry2D) register('ch.swisstopo.fixpunkte-hfp2', FixpunkteHfp2)
34.190083
60
0.673193
459
4,137
5.77342
0.150327
0.041509
0.075472
0.035849
0.809057
0.714717
0.714717
0.714717
0.695472
0.641132
0
0.031582
0.181049
4,137
120
61
34.475
0.75059
0.005076
0
0.712871
0
0
0.222654
0.106466
0
0
0
0
0
1
0
false
0
0.039604
0
0.940594
0
0
0
0
null
0
0
0
1
1
1
1
0
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
1
0
0
7
aa36a6fe7866049bd9b571b53e77325c1cd01652
741
py
Python
stubs/micropython-v1_13-95-pyboard/utime.py
mattytrentini/micropython-stubs
4d596273823b69e9e5bcf5fa67f249c374ee0bbc
[ "MIT" ]
null
null
null
stubs/micropython-v1_13-95-pyboard/utime.py
mattytrentini/micropython-stubs
4d596273823b69e9e5bcf5fa67f249c374ee0bbc
[ "MIT" ]
null
null
null
stubs/micropython-v1_13-95-pyboard/utime.py
mattytrentini/micropython-stubs
4d596273823b69e9e5bcf5fa67f249c374ee0bbc
[ "MIT" ]
null
null
null
""" Module: 'utime' on pyboard 1.13.0-95 """ # MCU: (sysname='pyboard', nodename='pyboard', release='1.13.0', version='v1.13-95-g0fff2e03f on 2020-10-03', machine='PYBv1.1 with STM32F405RG') # Stubber: 1.3.4 - updated from typing import Any def gmtime(*args) -> Any: pass def localtime(*args) -> Any: pass def mktime(*args) -> Any: pass def sleep(*args) -> Any: pass def sleep_ms(*args) -> Any: pass def sleep_us(*args) -> Any: pass def ticks_add(*args) -> Any: pass def ticks_cpu(*args) -> Any: pass def ticks_diff(*args) -> Any: pass def ticks_ms(*args) -> Any: pass def ticks_us(*args) -> Any: pass def time(*args) -> Any: pass def time_ns(*args) -> Any: pass
12.559322
145
0.60054
113
741
3.867257
0.39823
0.208238
0.327231
0.384439
0.448513
0
0
0
0
0
0
0.064799
0.22942
741
58
146
12.775862
0.700525
0.278003
0
0.481481
0
0
0
0
0
0
0
0
0
1
0.481481
true
0.481481
0.037037
0
0.518519
0
0
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
1
0
0
1
0
0
7
a4aaaaece86d0c63ed9292f64859cb26453869ba
3,806
py
Python
utility/refined_events/dialer_record.py
EfficientAI/efficient_cv
e308f229e4d99da86ad56f87f3a78b2c81f27ca5
[ "MIT" ]
null
null
null
utility/refined_events/dialer_record.py
EfficientAI/efficient_cv
e308f229e4d99da86ad56f87f3a78b2c81f27ca5
[ "MIT" ]
null
null
null
utility/refined_events/dialer_record.py
EfficientAI/efficient_cv
e308f229e4d99da86ad56f87f3a78b2c81f27ca5
[ "MIT" ]
null
null
null
from com.android.monkeyrunner import MonkeyRunner from com.android.monkeyrunner import MonkeyDevice print('Connecting to device...') device = MonkeyRunner.waitForConnection() print('Connected to device') # Reproduce action log from here print('Start to reproduce action log') device.touch(536, 1708, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(536, 1708, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(1.0) device.touch(148, 312, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(148, 312, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(1.0) device.touch(950, 1750, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(118, 608, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(1.0) device.touch(199, 1064, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(199, 1064, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(543, 1060, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(543, 1060, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(867, 1096, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(867, 1096, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(182, 1260, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(182, 1260, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(560, 1244, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(560, 1244, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(897, 1232, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(897, 1232, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(219, 1412, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(219, 1412, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(536, 1432, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(536, 1432, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(914, 1408, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(914, 1408, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(0.2) device.touch(992, 872, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(1.0) device.touch(816, 228, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(816, 228, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(1.0) device.touch(381, 656, MonkeyDevice.DOWN_AND_UP) print('Executing : device.touch(381, 656, MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(1.0) device.press("KEYCODE_HOME", MonkeyDevice.DOWN_AND_UP) print('Executing : device.press("KEYCODE_HOME", MonkeyDevice.DOWN_AND_UP)') MonkeyRunner.sleep(1.0) print('Finish to reproduce action log')
35.90566
75
0.788229
568
3,806
5.109155
0.107394
0.264645
0.314266
0.347347
0.929704
0.90765
0.90765
0.902826
0.616127
0.566506
0
0.097623
0.071466
3,806
106
76
35.90566
0.723543
0.007882
0
0.531646
0
0
0.415629
0.166358
0
0
0
0
0
1
0
false
0
0.025316
0
0.025316
0.35443
0
0
0
null
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
101f35eaf0e79f3f32b7e5bb7d3da1592de4065c
10,439
py
Python
dialogue-engine/test/programytest/parser/template/graph_tests/test_set.py
cotobadesign/cotoba-agent-oss
3833d56e79dcd7529c3e8b3a3a8a782d513d9b12
[ "MIT" ]
104
2020-03-30T09:40:00.000Z
2022-03-06T22:34:25.000Z
dialogue-engine/test/programytest/parser/template/graph_tests/test_set.py
cotobadesign/cotoba-agent-oss
3833d56e79dcd7529c3e8b3a3a8a782d513d9b12
[ "MIT" ]
25
2020-06-12T01:36:35.000Z
2022-02-19T07:30:44.000Z
dialogue-engine/test/programytest/parser/template/graph_tests/test_set.py
cotobadesign/cotoba-agent-oss
3833d56e79dcd7529c3e8b3a3a8a782d513d9b12
[ "MIT" ]
10
2020-04-02T23:43:56.000Z
2021-05-14T13:47:01.000Z
""" Copyright (c) 2020 COTOBA DESIGN, Inc. Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ import xml.etree.ElementTree as ET from programy.parser.template.nodes.base import TemplateNode from programy.parser.template.nodes.set import TemplateSetNode from programy.parser.exceptions import ParserException from programytest.parser.template.graph_tests.graph_test_client import TemplateGraphTestClient class TemplateGraphSetTests(TemplateGraphTestClient): def test_set_template_typename_as_attrib(self): template = ET.fromstring(""" <template> <set name="somepred">Value1</set> </template> """) ast = self._graph.parse_template_expression(template) self.assertIsNotNone(ast) self.assertIsInstance(ast, TemplateNode) self.assertIsNotNone(ast.children) self.assertEqual(len(ast.children), 1) set_node = ast.children[0] self.assertIsNotNone(set_node) self.assertIsInstance(set_node, TemplateSetNode) self.assertIsNotNone(set_node.name) self.assertIsInstance(set_node.name, TemplateNode) self.assertEqual(set_node.name.resolve(self._client_context), "somepred") self.assertEqual(set_node.property_type, "name") self.assertEqual(len(set_node.children), 1) self.assertEqual(set_node.children[0].resolve(self._client_context), "Value1") def test_set_template_multi_word_typename_as_attrib(self): template = ET.fromstring(""" <template> <set name="somepred other">Value1</set> </template> """) ast = self._graph.parse_template_expression(template) self.assertIsNotNone(ast) self.assertIsInstance(ast, TemplateNode) self.assertIsNotNone(ast.children) self.assertEqual(len(ast.children), 1) set_node = ast.children[0] self.assertIsNotNone(set_node) self.assertIsInstance(set_node, TemplateSetNode) self.assertIsNotNone(set_node.name) self.assertIsInstance(set_node.name, TemplateNode) self.assertEqual(set_node.name.resolve(self._client_context), "somepred other") self.assertEqual(set_node.property_type, "name") self.assertEqual(len(set_node.children), 1) self.assertEqual(set_node.children[0].resolve(self._client_context), "Value1") def test_set_template_typename_nested(self): template = ET.fromstring(""" <template> Some text here <set name="somepred">Value1</set> Some text there </template> """) ast = self._graph.parse_template_expression(template) self.assertIsNotNone(ast) self.assertIsInstance(ast, TemplateNode) self.assertIsNotNone(ast.children) self.assertEqual(len(ast.children), 3) set_node = ast.children[1] self.assertIsNotNone(set_node) self.assertIsInstance(set_node, TemplateSetNode) self.assertIsNotNone(set_node.name) self.assertIsInstance(set_node.name, TemplateNode) self.assertEqual(set_node.name.resolve(self._client_context), "somepred") self.assertEqual(set_node.property_type, "name") self.assertEqual(len(set_node.children), 1) self.assertEqual(set_node.children[0].resolve(self._client_context), "Value1") def test_set_template_typedata_as_attrib(self): template = ET.fromstring(""" <template> <set data="somedata">Value2</set> </template> """) ast = self._graph.parse_template_expression(template) self.assertIsNotNone(ast) self.assertIsInstance(ast, TemplateNode) self.assertIsNotNone(ast.children) self.assertEqual(len(ast.children), 1) set_node = ast.children[0] self.assertIsNotNone(set_node) self.assertIsInstance(set_node, TemplateSetNode) self.assertIsNotNone(set_node.name) self.assertIsInstance(set_node.name, TemplateNode) self.assertEqual(set_node.name.resolve(self._client_context), "somedata") self.assertEqual(set_node.property_type, "data") self.assertEqual(len(set_node.children), 1) self.assertEqual(set_node.children[0].resolve(self._client_context), "Value2") def test_set_template_typevar_as_attrib(self): template = ET.fromstring(""" <template> <set var="somevar">Value3</set> </template> """) ast = self._graph.parse_template_expression(template) self.assertIsNotNone(ast) self.assertIsInstance(ast, TemplateNode) self.assertIsNotNone(ast.children) self.assertEqual(len(ast.children), 1) set_node = ast.children[0] self.assertIsNotNone(set_node) self.assertIsInstance(set_node, TemplateSetNode) self.assertIsNotNone(set_node.name) self.assertIsInstance(set_node.name, TemplateNode) self.assertEqual(set_node.name.resolve(self._client_context), "somevar") self.assertEqual(set_node.property_type, "var") self.assertEqual(len(set_node.children), 1) self.assertEqual(set_node.children[0].resolve(self._client_context), "Value3") def test_set_template_typename_as_child(self): template = ET.fromstring(""" <template> <set><name>somepred</name>Value4</set> </template> """) ast = self._graph.parse_template_expression(template) self.assertIsNotNone(ast) self.assertIsInstance(ast, TemplateNode) self.assertIsNotNone(ast.children) self.assertEqual(len(ast.children), 1) set_node = ast.children[0] self.assertIsNotNone(set_node) self.assertIsInstance(set_node, TemplateSetNode) self.assertIsNotNone(set_node.name) self.assertIsInstance(set_node.name, TemplateNode) self.assertEqual(set_node.name.resolve(self._client_context), "somepred") self.assertEqual(set_node.property_type, "name") self.assertEqual(len(set_node.children), 1) self.assertEqual(set_node.children[0].resolve(self._client_context), "Value4") def test_set_template_typedata_as_child(self): template = ET.fromstring(""" <template> <set><data>somedata</data>Value5</set> </template> """) ast = self._graph.parse_template_expression(template) self.assertIsNotNone(ast) self.assertIsInstance(ast, TemplateNode) self.assertIsNotNone(ast.children) self.assertEqual(len(ast.children), 1) set_node = ast.children[0] self.assertIsNotNone(set_node) self.assertIsInstance(set_node, TemplateSetNode) self.assertIsNotNone(set_node.name) self.assertIsInstance(set_node.name, TemplateNode) self.assertEqual(set_node.name.resolve(self._client_context), "somedata") self.assertEqual(set_node.property_type, "data") self.assertEqual(len(set_node.children), 1) self.assertEqual(set_node.children[0].resolve(self._client_context), "Value5") def test_set_template_typevar_as_child(self): template = ET.fromstring(""" <template> <set><var>somevar</var>Value6</set> </template> """) ast = self._graph.parse_template_expression(template) self.assertIsNotNone(ast) self.assertIsInstance(ast, TemplateNode) self.assertIsNotNone(ast.children) self.assertEqual(len(ast.children), 1) set_node = ast.children[0] self.assertIsNotNone(set_node) self.assertIsInstance(set_node, TemplateSetNode) self.assertIsNotNone(set_node.name) self.assertIsInstance(set_node.name, TemplateNode) self.assertEqual(set_node.name.resolve(self._client_context), "somevar") self.assertEqual(set_node.property_type, "var") self.assertEqual(len(set_node.children), 1) self.assertEqual(set_node.children[0].resolve(self._client_context), "Value6") def test_set_type_name_and_var(self): template = ET.fromstring(""" <template> <set name="somepred" var="somevar">Value1</set> </template> """) with self.assertRaises(ParserException): self._graph.parse_template_expression(template) def test_set_type_name_and_data(self): template = ET.fromstring(""" <template> <set name="somepred" data="somedata">Value1</set> </template> """) with self.assertRaises(ParserException): self._graph.parse_template_expression(template) def test_set_type_data_and_var(self): template = ET.fromstring(""" <template> <set data="somedata" var="somevar">Value1</set> </template> """) with self.assertRaises(ParserException): self._graph.parse_template_expression(template) def test_set_other(self): template = ET.fromstring(""" <template> <set>Value1</set> </template> """) with self.assertRaises(ParserException): self._graph.parse_template_expression(template)
42.092742
126
0.666635
1,152
10,439
5.859375
0.141493
0.074667
0.039111
0.078222
0.810667
0.792
0.764148
0.764148
0.701778
0.701778
0
0.006978
0.231248
10,439
247
127
42.263158
0.834143
0.101734
0
0.81
0
0
0.174725
0.032127
0
0
0
0
0.5
1
0.06
false
0
0.025
0
0.09
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
1
0
0
0
0
0
0
0
0
0
8
103f2521cefdc18868ad041de3f7cb2d282e2f30
1,952
py
Python
heltour/tournament/migrations/0075_auto_20160818_0456.py
lenguyenthanh/heltour
13018b1905539de0b273370a76f6aa1d1ebbb01a
[ "MIT" ]
null
null
null
heltour/tournament/migrations/0075_auto_20160818_0456.py
lenguyenthanh/heltour
13018b1905539de0b273370a76f6aa1d1ebbb01a
[ "MIT" ]
null
null
null
heltour/tournament/migrations/0075_auto_20160818_0456.py
lenguyenthanh/heltour
13018b1905539de0b273370a76f6aa1d1ebbb01a
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- # Generated by Django 1.9.7 on 2016-08-18 04:56 from __future__ import unicode_literals from django.db import migrations import heltour.tournament.models class Migration(migrations.Migration): dependencies = [ ('tournament', '0074_auto_20160817_2247'), ] operations = [ migrations.AlterField( model_name='loneplayerscore', name='late_join_points', field=heltour.tournament.models.ScoreField(default=0), ), migrations.AlterField( model_name='loneplayerscore', name='points', field=heltour.tournament.models.ScoreField(default=0), ), migrations.AlterField( model_name='loneplayerscore', name='tiebreak1', field=heltour.tournament.models.ScoreField(default=0), ), migrations.AlterField( model_name='loneplayerscore', name='tiebreak2', field=heltour.tournament.models.ScoreField(default=0), ), migrations.AlterField( model_name='loneplayerscore', name='tiebreak3', field=heltour.tournament.models.ScoreField(default=0), ), migrations.AlterField( model_name='loneplayerscore', name='tiebreak4', field=heltour.tournament.models.ScoreField(default=0), ), migrations.AlterField( model_name='teampairing', name='black_points', field=heltour.tournament.models.ScoreField(default=0), ), migrations.AlterField( model_name='teampairing', name='white_points', field=heltour.tournament.models.ScoreField(default=0), ), migrations.AlterField( model_name='teamscore', name='game_points', field=heltour.tournament.models.ScoreField(default=0), ), ]
31.483871
66
0.59375
171
1,952
6.649123
0.309942
0.149516
0.202287
0.229551
0.746702
0.746702
0.704485
0.704485
0.658751
0.658751
0
0.032727
0.295594
1,952
61
67
32
0.794182
0.034324
0
0.648148
1
0
0.131243
0.012221
0
0
0
0
0
1
0
false
0
0.055556
0
0.111111
0
0
0
0
null
0
1
1
0
1
1
1
0
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
10487f6ac5c71017ff49947a3dbd1c852de25b8e
59
py
Python
src/ckanext-uebpackage/ckanext/uebpackage/celery_import.py
CI-WATER/portal
c61660c8389c7af82517cbd0154bc83f9737c4d1
[ "BSD-3-Clause" ]
1
2017-12-22T04:53:06.000Z
2017-12-22T04:53:06.000Z
src/ckanext-uebpackage/ckanext/uebpackage/celery_import.py
CI-WATER/portal
c61660c8389c7af82517cbd0154bc83f9737c4d1
[ "BSD-3-Clause" ]
4
2015-04-16T19:41:17.000Z
2020-10-25T18:02:14.000Z
src/ckanext-uebpackage/ckanext/uebpackage/celery_import.py
CI-WATER/portal
c61660c8389c7af82517cbd0154bc83f9737c4d1
[ "BSD-3-Clause" ]
null
null
null
def task_imports(): return ['ckanext.uebpackage.tasks']
29.5
39
0.728814
7
59
6
1
0
0
0
0
0
0
0
0
0
0
0
0.118644
59
2
39
29.5
0.807692
0
0
0
0
0
0.4
0.4
0
0
0
0
0
1
0.5
true
0
0.5
0.5
1.5
0
1
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
0
1
1
1
0
0
7
1085f313ac17a9f85f8d984b98c9da2c9b3da552
113
py
Python
pfdicom_tagSub/__init__.py
reddigari/pfdicom_tagSub
0eb8f7be164f6f59db35b18ab94b25b887b8163e
[ "Apache-2.0" ]
null
null
null
pfdicom_tagSub/__init__.py
reddigari/pfdicom_tagSub
0eb8f7be164f6f59db35b18ab94b25b887b8163e
[ "Apache-2.0" ]
1
2021-04-08T18:52:04.000Z
2021-04-08T18:52:04.000Z
pfdicom_tagSub/__init__.py
reddigari/pfdicom_tagSub
0eb8f7be164f6f59db35b18ab94b25b887b8163e
[ "Apache-2.0" ]
1
2020-12-03T18:09:37.000Z
2020-12-03T18:09:37.000Z
try: from .pfdicom_tagSub import pfdicom_tagSub except: from pfdicom_tagSub import pfdicom_tagSub
22.6
49
0.752212
14
113
5.785714
0.428571
0.641975
0.419753
0.567901
0.888889
0.888889
0
0
0
0
0
0
0.221239
113
4
50
28.25
0.920455
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
0.5
0
0.5
0
1
0
0
null
1
1
1
1
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
0
0
0
9
52cf61493d0499e13fff8039d5a57667cca9b210
47,656
py
Python
src/genie/libs/parser/iosxr/tests/test_show_igmp.py
nujo/genieparser
083b01efc46afc32abe1a1858729578beab50cd3
[ "Apache-2.0" ]
2
2021-01-27T03:37:39.000Z
2021-01-27T03:40:50.000Z
src/genie/libs/parser/iosxr/tests/test_show_igmp.py
nujo/genieparser
083b01efc46afc32abe1a1858729578beab50cd3
[ "Apache-2.0" ]
1
2020-08-01T00:23:31.000Z
2020-08-01T00:40:05.000Z
src/genie/libs/parser/iosxr/tests/test_show_igmp.py
nujo/genieparser
083b01efc46afc32abe1a1858729578beab50cd3
[ "Apache-2.0" ]
null
null
null
import re import unittest from unittest.mock import Mock from pyats.topology import Device from genie.metaparser.util.exceptions import SchemaEmptyParserError, \ SchemaMissingKeyError from genie.libs.parser.iosxr.show_igmp import ShowIgmpInterface, ShowIgmpSummary, ShowIgmpGroupsDetail ############################################################################# # unitest For Show IGMP Interface ############################################################################# class TestShowIgmpInterface(unittest.TestCase): device = Device(name='aDevice') device0 = Device(name='bDevice') empty_output = {'execute.return_value': ''} golden_parsed_output = { 'vrf': { 'default': { 'interfaces': { 'GigabitEthernet0/0/0/0.110': { 'interface_status': 'up', 'igmp_activity': { 'joins': 7, 'leaves': 1 }, 'igmp_max_query_response_time': 10, 'igmp_querier_timeout': 125, 'igmp_query_interval': 60, 'last_member_query_response_interval': 1, 'igmp_querying_router': '10.12.110.1', 'igmp_state': 'enabled', 'time_elapsed_since_last_query_sent': '02:42:58', 'time_elapsed_since_last_report_received': '00:00:31', 'time_elapsed_since_router_enabled': '02:46:41', 'igmp_version': 3, 'ip_address': '10.12.110.2/24', 'line_protocol': 'up', 'oper_status': 'up' }, 'GigabitEthernet0/0/0/0.115': { 'interface_status': 'up', 'igmp_state': 'disabled', 'ip_address': '10.12.115.2/24', 'line_protocol': 'up', 'oper_status': 'up' }, 'GigabitEthernet0/0/0/0.120': { 'interface_status': 'up', 'igmp_state': 'disabled', 'ip_address': '10.12.120.2/24', 'line_protocol': 'up', 'oper_status': 'up' }, 'GigabitEthernet0/0/0/0.90':{ 'interface_status': 'up', 'igmp_state': 'disabled', 'ip_address': '10.12.90.2/24', 'line_protocol': 'up', 'oper_status': 'up' }, 'GigabitEthernet0/0/0/1.110': { 'interface_status': 'up', 'igmp_activity': { 'joins': 5, 'leaves': 0 }, 'igmp_max_query_response_time': 10, 'igmp_querier_timeout': 125, 'igmp_query_interval': 60, 'last_member_query_response_interval': 1, 'igmp_querying_router': '10.23.110.2', 'igmp_querying_router_info': 'this system', 'igmp_state': 'enabled', 'time_elapsed_since_last_query_sent': '00:00:55', 'time_elapsed_since_last_report_received': '00:00:55', 'time_elapsed_since_router_enabled': '02:46:41', 'igmp_version': 3, 'ip_address': '10.23.110.2/24', 'line_protocol': 'up', 'oper_status': 'up' }, 'GigabitEthernet0/0/0/1.115': { 'interface_status': 'up', 'igmp_state': 'disabled', 'ip_address': '10.23.115.2/24', 'line_protocol': 'up', 'oper_status': 'up' }, 'GigabitEthernet0/0/0/1.120': { 'interface_status': 'up', 'igmp_state': 'disabled', 'ip_address': '10.23.120.2/24', 'line_protocol': 'up', 'oper_status': 'up' }, 'GigabitEthernet0/0/0/1.90': { 'interface_status': 'up', 'igmp_state': 'disabled', 'ip_address': '10.23.90.2/24', 'line_protocol': 'up', 'oper_status': 'up' }, 'Loopback0': { 'interface_status': 'up', 'igmp_activity': { 'joins': 6, 'leaves': 0 }, 'igmp_max_query_response_time': 10, 'igmp_querier_timeout': 125, 'igmp_query_interval': 60, 'last_member_query_response_interval': 1, 'igmp_querying_router': '10.16.2.2', 'igmp_querying_router_info': 'this system', 'igmp_state': 'enabled', 'time_elapsed_since_last_query_sent': '00:00:53', 'time_elapsed_since_last_report_received': '00:00:51', 'time_elapsed_since_router_enabled': '02:46:41', 'igmp_version': 3, 'ip_address': '10.16.2.2/32', 'line_protocol': 'up', 'oper_status': 'up' } } } } } golden_output = {'execute.return_value': ''' R2_xr# show igmp interface Loopback0 is up, line protocol is up Internet address is 10.16.2.2/32 IGMP is enabled on interface Current IGMP version is 3 IGMP query interval is 60 seconds IGMP querier timeout is 125 seconds IGMP max query response time is 10 seconds Last member query response interval is 1 seconds IGMP activity: 6 joins, 0 leaves IGMP querying router is 10.16.2.2 (this system) Time elapsed since last query sent 00:00:53 Time elapsed since IGMP router enabled 02:46:41 Time elapsed since last report received 00:00:51 GigabitEthernet0/0/0/0.90 is up, line protocol is up Internet address is 10.12.90.2/24 IGMP is disabled on interface GigabitEthernet0/0/0/1.90 is up, line protocol is up Internet address is 10.23.90.2/24 IGMP is disabled on interface GigabitEthernet0/0/0/0.110 is up, line protocol is up Internet address is 10.12.110.2/24 IGMP is enabled on interface Current IGMP version is 3 IGMP query interval is 60 seconds IGMP querier timeout is 125 seconds IGMP max query response time is 10 seconds Last member query response interval is 1 seconds IGMP activity: 7 joins, 1 leaves IGMP querying router is 10.12.110.1 Time elapsed since last query sent 02:42:58 Time elapsed since IGMP router enabled 02:46:41 Time elapsed since last report received 00:00:31 GigabitEthernet0/0/0/0.115 is up, line protocol is up Internet address is 10.12.115.2/24 IGMP is disabled on interface GigabitEthernet0/0/0/0.120 is up, line protocol is up Internet address is 10.12.120.2/24 IGMP is disabled on interface GigabitEthernet0/0/0/1.110 is up, line protocol is up Internet address is 10.23.110.2/24 IGMP is enabled on interface Current IGMP version is 3 IGMP query interval is 60 seconds IGMP querier timeout is 125 seconds IGMP max query response time is 10 seconds Last member query response interval is 1 seconds IGMP activity: 5 joins, 0 leaves IGMP querying router is 10.23.110.2 (this system) Time elapsed since last query sent 00:00:55 Time elapsed since IGMP router enabled 02:46:41 Time elapsed since last report received 00:00:55 GigabitEthernet0/0/0/1.115 is up, line protocol is up Internet address is 10.23.115.2/24 IGMP is disabled on interface GigabitEthernet0/0/0/1.120 is up, line protocol is up Internet address is 10.23.120.2/24 IGMP is disabled on interface '''} golden_parsed_interface_output = { 'vrf': { 'default': { 'interfaces': { 'GigabitEthernet0/0/0/1.115': { 'interface_status': 'up', 'igmp_state': 'disabled', 'ip_address': '10.23.115.2/24', 'line_protocol': 'up', 'oper_status': 'up' } } } } } golden_parsed_interface_output1 = { 'vrf': { 'VRF1': { 'interfaces': { 'Loopback0': { 'interface_status': 'up', 'igmp_activity': { 'joins': 6, 'leaves': 0 }, 'igmp_max_query_response_time': 10, 'igmp_querier_timeout': 125, 'igmp_query_interval': 60, 'last_member_query_response_interval': 1, 'igmp_querying_router': '10.16.2.2', 'igmp_querying_router_info': 'this system', 'igmp_state': 'enabled', 'time_elapsed_since_last_query_sent': '00:00:53', 'time_elapsed_since_last_report_received': '00:00:51', 'time_elapsed_since_router_enabled': '02:46:41', 'igmp_version': 3, 'ip_address': '10.16.2.2/32', 'line_protocol': 'up', 'oper_status': 'up' } } } } } golden_interface_output={'execute.return_value':''' R2_xr# show igmp interface GigabitEthernet0/0/0/1.115 GigabitEthernet0/0/0/1.115 is up, line protocol is up Internet address is 10.23.115.2/24 IGMP is disabled on interface '''} golden_interface_output1={'execute.return_value':''' R2_xr# show igmp vrf VRF1 interface Loopback0 is up, line protocol is up Internet address is 10.16.2.2/32 IGMP is enabled on interface Current IGMP version is 3 IGMP query interval is 60 seconds IGMP querier timeout is 125 seconds IGMP max query response time is 10 seconds Last member query response interval is 1 seconds IGMP activity: 6 joins, 0 leaves IGMP querying router is 10.16.2.2 (this system) Time elapsed since last query sent 00:00:53 Time elapsed since IGMP router enabled 02:46:41 Time elapsed since last report received 00:00:51 '''} golden_parsed_interface_output2 = { 'vrf': { 'VRF1': { 'interfaces': { 'Loopback0': { 'interface_status': 'up', 'igmp_activity': { 'joins': 6, 'leaves': 0 }, 'igmp_max_query_response_time': 10, 'igmp_querier_timeout': 125, 'igmp_query_interval': 60, 'last_member_query_response_interval': 1, 'igmp_querying_router': '10.16.2.2', 'igmp_querying_router_info': 'this system', 'igmp_state': 'enabled', 'time_elapsed_since_last_query_sent': '00:00:53', 'time_elapsed_since_last_report_received': '00:00:51', 'time_elapsed_since_router_enabled': '02:46:41', 'igmp_version': 3, 'ip_address': '10.16.2.2/32', 'line_protocol': 'up', 'oper_status': 'up' } } } } } golden_interface_output2={'execute.return_value':''' R2_xr# show igmp vrf VRF1 interface Loopback0 Loopback0 is up, line protocol is up Internet address is 10.16.2.2/32 IGMP is enabled on interface Current IGMP version is 3 IGMP query interval is 60 seconds IGMP querier timeout is 125 seconds IGMP max query response time is 10 seconds Last member query response interval is 1 seconds IGMP activity: 6 joins, 0 leaves IGMP querying router is 10.16.2.2 (this system) Time elapsed since last query sent 00:00:53 Time elapsed since IGMP router enabled 02:46:41 Time elapsed since last report received 00:00:51 '''} def test_empty(self): self.device1 = Mock(**self.empty_output) interface_detail_obj = ShowIgmpInterface(device=self.device1) with self.assertRaises(SchemaEmptyParserError): parsed_output = interface_detail_obj.parse() def test_golden(self): self.device = Mock(**self.golden_output) interface_detail_obj = ShowIgmpInterface(device=self.device) parsed_output = interface_detail_obj.parse() self.maxDiff = None self.assertEqual(parsed_output,self.golden_parsed_output) def test_golden_custom_interface(self): self.device = Mock(**self.golden_interface_output) interface_detail_obj = ShowIgmpInterface(device=self.device) parsed_output = interface_detail_obj.parse(interface='GigabitEthernet0/0/0/1.115') self.maxDiff = None self.assertEqual(parsed_output, self.golden_parsed_interface_output) def test_golden_custom_vrf(self): self.device = Mock(**self.golden_interface_output1) interface_detail_obj = ShowIgmpInterface(device=self.device) parsed_output = interface_detail_obj.parse(vrf='VRF1') self.maxDiff = None self.assertEqual(parsed_output, self.golden_parsed_interface_output1) def test_golden_custom_vrf_interface(self): self.device = Mock(**self.golden_interface_output2) interface_detail_obj = ShowIgmpInterface(device=self.device) parsed_output = interface_detail_obj.parse(vrf='VRF1', interface='Loopback0') self.maxDiff = None self.assertEqual(parsed_output, self.golden_parsed_interface_output2) ############################################################################# # unitest For Show IGMP Summary ############################################################################# class test_show_igmp_summary(unittest.TestCase): device = Device(name='aDevice') device0 = Device(name='bDevice') empty_output = {'execute.return_value': ''} golden_parsed_output = { 'vrf': { 'default': { 'disabled_interfaces': 6, 'enabled_interfaces': 3, 'no_of_group_x_interface': 16, 'interfaces': { 'Loopback0': { 'max_groups': 25000, 'number_groups': 6 }, 'GigabitEthernet0/0/0/0.90': { 'max_groups': 25000, 'number_groups': 1 }, 'GigabitEthernet0/0/0/1.90': { 'max_groups': 25000, 'number_groups': 1 }, 'GigabitEthernet0/0/0/0.110': { 'max_groups': 25000, 'number_groups': 6 }, 'GigabitEthernet0/0/0/0.115': { 'max_groups': 25000, 'number_groups': 4 }, 'GigabitEthernet0/0/0/0.120': { 'max_groups': 25000, 'number_groups': 1 }, 'GigabitEthernet0/0/0/1.110': { 'max_groups': 25000, 'number_groups': 5 }, 'GigabitEthernet0/0/0/1.115': { 'max_groups': 25000, 'number_groups': 0 }, 'GigabitEthernet0/0/0/1.120': { 'max_groups': 25000, 'number_groups': 1 } }, 'mte_tuple_count': 0, 'maximum_number_of_groups_for_vrf': 50000, 'robustness_value': 2, 'supported_interfaces': 9, 'unsupported_interfaces': 0, } } } golden_output = {'execute.return_value': ''' R2_xr#show igmp summary Robustness Value 2 No. of Group x Interfaces 16 Maximum number of Groups for this VRF 50000 Supported Interfaces : 9 Unsupported Interfaces : 0 Enabled Interfaces : 3 Disabled Interfaces : 6 MTE tuple count : 0 Interface Number Max # Groups Groups Loopback0 6 25000 GigabitEthernet0/0/0/0.90 1 25000 GigabitEthernet0/0/0/1.90 1 25000 GigabitEthernet0/0/0/0.110 6 25000 GigabitEthernet0/0/0/0.115 4 25000 GigabitEthernet0/0/0/0.120 1 25000 GigabitEthernet0/0/0/1.110 5 25000 GigabitEthernet0/0/0/1.115 0 25000 GigabitEthernet0/0/0/1.120 1 25000 '''} golden_parsed_summary_output = { 'vrf': { 'VRF1': { 'disabled_interfaces': 6, 'enabled_interfaces': 3, 'no_of_group_x_interface': 15, 'interfaces': { 'Loopback300': { 'max_groups': 25000, 'number_groups': 4 }, 'GigabitEthernet0/0/0/0.390': { 'max_groups': 25000, 'number_groups': 1 }, 'GigabitEthernet0/0/0/0.410': { 'max_groups': 25000, 'number_groups': 7 }, 'GigabitEthernet0/0/0/0.415': { 'max_groups': 25000, 'number_groups': 4 }, 'GigabitEthernet0/0/0/0.420': { 'max_groups': 25000, 'number_groups': 1 }, 'GigabitEthernet0/0/0/1.390': { 'max_groups': 25000, 'number_groups': 1 }, 'GigabitEthernet0/0/0/1.410': { 'max_groups': 25000, 'number_groups': 5 }, 'GigabitEthernet0/0/0/1.415': { 'max_groups': 25000, 'number_groups': 0 }, 'GigabitEthernet0/0/0/1.420': { 'max_groups': 25000, 'number_groups': 1 } }, 'mte_tuple_count': 0, 'maximum_number_of_groups_for_vrf': 50000, 'robustness_value': 2, 'supported_interfaces': 9, 'unsupported_interfaces': 0, } } } golden_summary_output={'execute.return_value':''' R2_xr#show igmp vrf VRF1 summary Robustness Value 2 No. of Group x Interfaces 15 Maximum number of Groups for this VRF 50000 Supported Interfaces : 9 Unsupported Interfaces : 0 Enabled Interfaces : 3 Disabled Interfaces : 6 MTE tuple count : 0 Interface Number Max # Groups Groups Loopback300 4 25000 GigabitEthernet0/0/0/0.390 1 25000 GigabitEthernet0/0/0/0.410 7 25000 GigabitEthernet0/0/0/0.415 4 25000 GigabitEthernet0/0/0/0.420 1 25000 GigabitEthernet0/0/0/1.390 1 25000 GigabitEthernet0/0/0/1.410 5 25000 GigabitEthernet0/0/0/1.415 0 25000 GigabitEthernet0/0/0/1.420 1 25000 '''} def test_empty(self): self.device1 = Mock(**self.empty_output) summary_detail_obj = ShowIgmpSummary(device=self.device1) with self.assertRaises(SchemaEmptyParserError): parsed_output = summary_detail_obj.parse() def test_golden(self): self.device = Mock(**self.golden_output) summary_detail_obj = ShowIgmpSummary(device=self.device) parsed_output = summary_detail_obj.parse() self.maxDiff = None self.assertEqual(parsed_output,self.golden_parsed_output) def test_golden_custom(self): self.device = Mock(**self.golden_summary_output) summary_detail_obj = ShowIgmpSummary(device=self.device) parsed_output = summary_detail_obj.parse(vrf='VRF1') self.maxDiff = None self.assertEqual(parsed_output, self.golden_parsed_summary_output) ############################################################################# # unitest For Show IGMP Groups Detail ############################################################################# class test_show_igmp_groups_detail(unittest.TestCase): device = Device(name='aDevice') device0 = Device(name='bDevice') empty_output = {'execute.return_value': ''} golden_parsed_output = { "vrf": { "default": { "interfaces": { "Loopback0": { "group": { "224.0.0.2": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:44:55" }, "224.0.0.9": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "09:47:23" }, "224.0.0.13": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:44:55" }, "224.0.0.22": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:44:55" }, "224.0.1.39": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:19:56" }, "224.0.1.40": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:44:55" } } }, "GigabitEthernet0/0/0/0.90": { "group": { "224.0.0.10": { "host_mode": "exclude", "last_reporter": "0.0.0.0", "router_mode": "INCLUDE", "router_mode_expires": "None", "suppress": 0, "up_time": "01:53:32" } } }, "GigabitEthernet0/0/0/0.110": { "group": { "224.0.0.2": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:44:55" }, "224.0.0.5": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "10:36:57" }, "224.0.0.6": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "10:36:57" }, "224.0.0.13": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:44:55" }, "224.0.0.22": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:44:55" }, "224.0.1.39": { "host_mode": "include", "last_reporter": "10.12.110.1", "router_mode": "EXCLUDE", "router_mode_expires": "00:01:41", "suppress": 0, "up_time": "02:29:47", } } }, "GigabitEthernet0/0/0/0.120": { "group": { "224.0.0.9": { "host_mode": "exclude", "last_reporter": "0.0.0.0", "router_mode": "INCLUDE", "router_mode_expires": "None", "suppress": 0, "up_time": "09:47:23" } } }, "GigabitEthernet0/0/0/1.120": { "group": { "224.0.0.9": { "host_mode": "exclude", "last_reporter": "0.0.0.0", "router_mode": "INCLUDE", "router_mode_expires": "None", "suppress": 0, "up_time": "09:47:23" } } } } } } } golden_output = {'execute.return_value': ''' R2_xr#show igmp groups detail Interface: Loopback0 Group: 224.0.0.2 Uptime: 02:44:55 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: Loopback0 Group: 224.0.0.9 Uptime: 09:47:23 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: Loopback0 Group: 224.0.0.13 Uptime: 02:44:55 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: Loopback0 Group: 224.0.0.22 Uptime: 02:44:55 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: Loopback0 Group: 224.0.1.39 Uptime: 02:19:56 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: Loopback0 Group: 224.0.1.40 Uptime: 02:44:55 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.90 Group: 224.0.0.10 Uptime: 01:53:32 Router mode: INCLUDE Host mode: EXCLUDE Last reporter: 0.0.0.0 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.110 Group: 224.0.0.2 Uptime: 02:44:55 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.110 Group: 224.0.0.5 Uptime: 10:36:57 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.110 Group: 224.0.0.6 Uptime: 10:36:57 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.110 Group: 224.0.0.13 Uptime: 02:44:55 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.110 Group: 224.0.0.22 Uptime: 02:44:55 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.110 Group: 224.0.1.39 Uptime: 02:29:47 Router mode: EXCLUDE (Expires: 00:01:41) Host mode: INCLUDE Last reporter: 10.12.110.1 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.120 Group: 224.0.0.9 Uptime: 09:47:23 Router mode: INCLUDE Host mode: EXCLUDE Last reporter: 0.0.0.0 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/1.120 Group: 224.0.0.9 Uptime: 09:47:23 Router mode: INCLUDE Host mode: EXCLUDE Last reporter: 0.0.0.0 Suppress: 0 Source list is empty '''} golden_parsed_igmp_groups_output = { "vrf": { "VRF1": { "interfaces": { "Loopback300": { "group": { "224.0.0.2": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:43:30" }, "224.0.0.9": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "09:48:07" }, "224.0.0.13": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:43:30" }, "224.0.0.22": { "host_mode": "exclude", "last_reporter": "10.16.2.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:43:30" } } }, "GigabitEthernet0/0/0/0.390": { "group": { "224.0.0.10": { "host_mode": "exclude", "last_reporter": "0.0.0.0", "router_mode": "INCLUDE", "router_mode_expires": "None", "suppress": 0, "up_time": "01:54:16" } } }, "GigabitEthernet0/0/0/0.410": { "group": { "224.0.0.2": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:43:30" }, "224.0.0.5": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "10:37:41" }, "224.0.0.6": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "10:37:41" }, "224.0.0.13": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:43:30" }, "224.0.0.22": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:43:30" }, "224.0.1.39": { "host_mode": "include", "last_reporter": "10.12.110.1", "router_mode": "EXCLUDE", "router_mode_expires": "00:01:21", "suppress": 0, "up_time": "02:30:06" }, "224.0.1.40": { "host_mode": "exclude", "last_reporter": "10.12.110.2", "router_mode": "EXCLUDE", "router_mode_expires": "never", "suppress": 0, "up_time": "02:43:30" } } }, "GigabitEthernet0/0/0/0.420": { "group": { "224.0.0.9": { "host_mode": "exclude", "last_reporter": "0.0.0.0", "router_mode": "INCLUDE", "router_mode_expires": "None", "suppress": 0, "up_time": "09:48:07" } } }, "GigabitEthernet0/0/0/1.390": { "group": { "224.0.0.10": { "host_mode": "exclude", "last_reporter": "0.0.0.0", "router_mode": "INCLUDE", "router_mode_expires": "None", "suppress": 0, "up_time": "01:54:16" } } }, "GigabitEthernet0/0/0/1.420": { "group": { "224.0.0.9": { "host_mode": "exclude", "last_reporter": "0.0.0.0", "router_mode": "INCLUDE", "router_mode_expires": "None", "suppress": 0, "up_time": "09:48:07" } } } } } } } golden_igmp_groups_output={'execute.return_value':''' R2_xr#show igmp vrf VRF1 groups detail Interface: Loopback300 Group: 224.0.0.2 Uptime: 02:43:30 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: Loopback300 Group: 224.0.0.9 Uptime: 09:48:07 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: Loopback300 Group: 224.0.0.13 Uptime: 02:43:30 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: Loopback300 Group: 224.0.0.22 Uptime: 02:43:30 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.16.2.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.390 Group: 224.0.0.10 Uptime: 01:54:16 Router mode: INCLUDE Host mode: EXCLUDE Last reporter: 0.0.0.0 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.410 Group: 224.0.0.2 Uptime: 02:43:30 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.410 Group: 224.0.0.5 Uptime: 10:37:41 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.410 Group: 224.0.0.6 Uptime: 10:37:41 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.410 Group: 224.0.0.13 Uptime: 02:43:30 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.410 Group: 224.0.0.22 Uptime: 02:43:30 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.410 Group: 224.0.1.39 Uptime: 02:30:06 Router mode: EXCLUDE (Expires: 00:01:21) Host mode: INCLUDE Last reporter: 10.12.110.1 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.410 Group: 224.0.1.40 Uptime: 02:43:30 Router mode: EXCLUDE (Expires: never) Host mode: EXCLUDE Last reporter: 10.12.110.2 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/0.420 Group: 224.0.0.9 Uptime: 09:48:07 Router mode: INCLUDE Host mode: EXCLUDE Last reporter: 0.0.0.0 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/1.390 Group: 224.0.0.10 Uptime: 01:54:16 Router mode: INCLUDE Host mode: EXCLUDE Last reporter: 0.0.0.0 Suppress: 0 Source list is empty Interface: GigabitEthernet0/0/0/1.420 Group: 224.0.0.9 Uptime: 09:48:07 Router mode: INCLUDE Host mode: EXCLUDE Last reporter: 0.0.0.0 Suppress: 0 Source list is empty '''} golden_output1 = {'execute.return_value': ''' R2_xr#show igmp groups detail Interface: GigabitEthernet0/0/0/2 Group: 232.1.1.1 Uptime: 00:04:55 Router mode: INCLUDE Host mode: INCLUDE Last reporter: 192.168.1.42 Group source list: Source Address Uptime Expires Fwd Flags 192.168.1.18 00:04:55 00:01:28 Yes Remote '''} golden_parsed_output1 = { "vrf": { "default": { "interfaces": { "GigabitEthernet0/0/0/2": { "group": { "232.1.1.1": { "host_mode": "include", "last_reporter": "192.168.1.42", "router_mode": "INCLUDE", "router_mode_expires": "None", "up_time": "00:04:55", "source": { "192.168.1.18": { "up_time": "00:04:55", "expire": "00:01:28", "forward": "Yes", "flags": "Remote" } } } } } } } } } def test_empty(self): self.device1 = Mock(**self.empty_output) igmp_groups_detail_obj = ShowIgmpGroupsDetail(device=self.device1) with self.assertRaises(SchemaEmptyParserError): parsed_output = igmp_groups_detail_obj.parse() def test_golden(self): self.device = Mock(**self.golden_output) igmp_groups_detail_obj = ShowIgmpGroupsDetail(device=self.device) parsed_output = igmp_groups_detail_obj.parse() self.maxDiff = None self.assertEqual(parsed_output,self.golden_parsed_output) def test_golden_custom(self): self.device = Mock(**self.golden_igmp_groups_output) igmp_groups_detail_obj = ShowIgmpGroupsDetail(device=self.device) parsed_output = igmp_groups_detail_obj.parse(vrf='VRF1') self.maxDiff = None self.assertEqual(parsed_output, self.golden_parsed_igmp_groups_output) def test_golden1(self): self.device = Mock(**self.golden_output1) igmp_groups_detail_obj = ShowIgmpGroupsDetail(device=self.device) parsed_output = igmp_groups_detail_obj.parse() self.maxDiff = None self.assertEqual(parsed_output,self.golden_parsed_output1) if __name__ == '__main__': unittest.main()
41.047373
102
0.419444
4,396
47,656
4.388763
0.045723
0.023014
0.077437
0.05515
0.94739
0.932359
0.896854
0.879127
0.867465
0.833204
0
0.112432
0.477422
47,656
1,160
103
41.082759
0.662263
0.002035
0
0.729656
0
0
0.486452
0.06878
0
0
0
0
0.01085
1
0.01085
false
0
0.005425
0
0.0434
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
52f9e2c4f204a43900b04f8eb2b268c771a17bf9
36
py
Python
ex1_5_execfile/simple.py
demsheng/QtBoostPythonHelloWorld
fae98d11c009495cc3b060a1980a0463652768da
[ "MIT" ]
null
null
null
ex1_5_execfile/simple.py
demsheng/QtBoostPythonHelloWorld
fae98d11c009495cc3b060a1980a0463652768da
[ "MIT" ]
null
null
null
ex1_5_execfile/simple.py
demsheng/QtBoostPythonHelloWorld
fae98d11c009495cc3b060a1980a0463652768da
[ "MIT" ]
null
null
null
def foo(i = 4): return i + 2008
12
19
0.527778
7
36
2.714286
0.857143
0
0
0
0
0
0
0
0
0
0
0.208333
0.333333
36
2
20
18
0.583333
0
0
0
0
0
0
0
0
0
0
0
0
1
0.5
false
0
0
0.5
1
0
1
1
0
null
0
0
0
0
0
0
0
0
0
0
1
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
0
1
1
0
0
7
eaad8e76c5419ad245a5e376f9ee44485551f19c
45,240
py
Python
tsupy/go.py
DataIntelligenceCrew/tsupy
363ecf5616ce4c2367b5550e6351548053a71bb6
[ "MIT" ]
1
2022-03-25T03:19:17.000Z
2022-03-25T03:19:17.000Z
tsupy/go.py
DataIntelligenceCrew/tsupy
363ecf5616ce4c2367b5550e6351548053a71bb6
[ "MIT" ]
null
null
null
tsupy/go.py
DataIntelligenceCrew/tsupy
363ecf5616ce4c2367b5550e6351548053a71bb6
[ "MIT" ]
2
2022-03-25T03:19:21.000Z
2022-03-30T04:02:55.000Z
# python wrapper for package go within overall package tsubasa # This is what you import to use the package. # File is generated by gopy. Do not edit. # gopy gen -output=tsupy -vm=python3 github.com/js061/tsubasa # # the following is required to enable dlopen to open the _go.so file import os,sys,inspect,collections try: import collections.abc as _collections_abc except ImportError: _collections_abc = collections cwd = os.getcwd() currentdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe()))) os.chdir(currentdir) from . import _tsubasa os.chdir(cwd) # to use this code in your end-user python file, import it as follows: # from tsubasa import go # and then refer to everything using go. prefix # packages imported by this package listed below: import collections try: import collections.abc as _collections_abc except ImportError: _collections_abc = collections class GoClass(object): """GoClass is the base class for all GoPy wrapper classes""" def __init__(self): self.handle = 0 # use go.nil for nil pointers nil = GoClass() # need to explicitly initialize it def main(): global nil nil = GoClass() main() def Init(): """calls the GoPyInit function, which runs the 'main' code string that was passed using -main arg to gopy""" _tsubasa.GoPyInit() # ---- Types --- # Python type for slice []bool class Slice_bool(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_bool_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_bool.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_bool len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_bool([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_bool_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_bool_len(self.handle) return Slice_bool(handle=_tsubasa.Slice_bool_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_bool_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_bool_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_bool.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_bool_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_bool_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []byte class Slice_byte(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_byte_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_byte.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_byte len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_byte([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_byte_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_byte_len(self.handle) return Slice_byte(handle=_tsubasa.Slice_byte_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_byte_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_byte_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_byte.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_byte_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_byte_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []float32 class Slice_float32(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_float32_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_float32.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_float32 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_float32([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_float32_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_float32_len(self.handle) return Slice_float32(handle=_tsubasa.Slice_float32_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_float32_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_float32_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_float32.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_float32_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_float32_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []float64 class Slice_float64(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_float64_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_float64.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_float64 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_float64([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_float64_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_float64_len(self.handle) return Slice_float64(handle=_tsubasa.Slice_float64_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_float64_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_float64_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_float64.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_float64_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_float64_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []int class Slice_int(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_int_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_int.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_int len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_int([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_int_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_int_len(self.handle) return Slice_int(handle=_tsubasa.Slice_int_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_int_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_int_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_int.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_int_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_int_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []int16 class Slice_int16(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_int16_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_int16.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_int16 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_int16([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_int16_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_int16_len(self.handle) return Slice_int16(handle=_tsubasa.Slice_int16_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_int16_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_int16_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_int16.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_int16_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_int16_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []int32 class Slice_int32(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_int32_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_int32.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_int32 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_int32([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_int32_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_int32_len(self.handle) return Slice_int32(handle=_tsubasa.Slice_int32_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_int32_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_int32_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_int32.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_int32_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_int32_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []int64 class Slice_int64(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_int64_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_int64.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_int64 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_int64([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_int64_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_int64_len(self.handle) return Slice_int64(handle=_tsubasa.Slice_int64_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_int64_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_int64_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_int64.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_int64_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_int64_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []int8 class Slice_int8(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_int8_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_int8.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_int8 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_int8([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_int8_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_int8_len(self.handle) return Slice_int8(handle=_tsubasa.Slice_int8_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_int8_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_int8_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_int8.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_int8_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_int8_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []rune class Slice_rune(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_rune_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_rune.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_rune len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_rune([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_rune_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_rune_len(self.handle) return Slice_rune(handle=_tsubasa.Slice_rune_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_rune_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_rune_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_rune.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_rune_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_rune_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []string class Slice_string(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_string_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_string.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_string len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_string([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_string_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_string_len(self.handle) return Slice_string(handle=_tsubasa.Slice_string_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_string_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_string_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_string.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_string_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_string_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []uint class Slice_uint(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_uint_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_uint.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_uint len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_uint([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_uint_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_uint_len(self.handle) return Slice_uint(handle=_tsubasa.Slice_uint_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_uint_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_uint_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_uint.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_uint_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_uint_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []uint16 class Slice_uint16(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_uint16_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_uint16.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_uint16 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_uint16([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_uint16_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_uint16_len(self.handle) return Slice_uint16(handle=_tsubasa.Slice_uint16_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_uint16_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_uint16_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_uint16.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_uint16_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_uint16_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []uint32 class Slice_uint32(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_uint32_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_uint32.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_uint32 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_uint32([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_uint32_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_uint32_len(self.handle) return Slice_uint32(handle=_tsubasa.Slice_uint32_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_uint32_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_uint32_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_uint32.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_uint32_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_uint32_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []uint64 class Slice_uint64(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_uint64_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_uint64.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_uint64 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_uint64([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_uint64_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_uint64_len(self.handle) return Slice_uint64(handle=_tsubasa.Slice_uint64_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_uint64_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_uint64_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_uint64.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_uint64_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_uint64_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # Python type for slice []uint8 class Slice_uint8(GoClass): """""" def __init__(self, *args, **kwargs): """ handle=A Go-side object is always initialized with an explicit handle=arg otherwise parameter is a python list that we copy from """ self.index = 0 if len(kwargs) == 1 and 'handle' in kwargs: self.handle = kwargs['handle'] _tsubasa.IncRef(self.handle) elif len(args) == 1 and isinstance(args[0], GoClass): self.handle = args[0].handle _tsubasa.IncRef(self.handle) else: self.handle = _tsubasa.Slice_uint8_CTor() _tsubasa.IncRef(self.handle) if len(args) > 0: if not isinstance(args[0], _collections_abc.Iterable): raise TypeError('Slice_uint8.__init__ takes a sequence as argument') for elt in args[0]: self.append(elt) def __del__(self): _tsubasa.DecRef(self.handle) def __str__(self): s = 'go.Slice_uint8 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' [' if len(self) < 120: s += ', '.join(map(str, self)) + ']' return s def __repr__(self): return 'go.Slice_uint8([' + ', '.join(map(str, self)) + '])' def __len__(self): return _tsubasa.Slice_uint8_len(self.handle) def __getitem__(self, key): if isinstance(key, slice): if key.step == None or key.step == 1: st = key.start ed = key.stop if st == None: st = 0 if ed == None: ed = _tsubasa.Slice_uint8_len(self.handle) return Slice_uint8(handle=_tsubasa.Slice_uint8_subslice(self.handle, st, ed)) return [self[ii] for ii in range(*key.indices(len(self)))] elif isinstance(key, int): if key < 0: key += len(self) if key < 0 or key >= len(self): raise IndexError('slice index out of range') return _tsubasa.Slice_uint8_elem(self.handle, key) else: raise TypeError('slice index invalid type') def __setitem__(self, idx, value): if idx < 0: idx += len(self) if idx < len(self): _tsubasa.Slice_uint8_set(self.handle, idx, value) return raise IndexError('slice index out of range') def __iadd__(self, value): if not isinstance(value, _collections_abc.Iterable): raise TypeError('Slice_uint8.__iadd__ takes a sequence as argument') for elt in value: self.append(elt) return self def __iter__(self): self.index = 0 return self def __next__(self): if self.index < len(self): rv = _tsubasa.Slice_uint8_elem(self.handle, self.index) self.index = self.index + 1 return rv raise StopIteration def append(self, value): _tsubasa.Slice_uint8_append(self.handle, value) def copy(self, src): """ copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """ mx = min(len(self), len(src)) for i in range(mx): self[i] = src[i] # ---- External Types Outside of Targeted Packages ---
32.59366
126
0.676415
6,844
45,240
4.267972
0.029369
0.087984
0.027936
0.037795
0.930503
0.924478
0.906265
0.900719
0.900719
0.900719
0
0.015391
0.191424
45,240
1,387
127
32.617159
0.783133
0.114478
0
0.817959
1
0
0.095411
0.006462
0
0
0
0
0
1
0.159184
false
0
0.005714
0.026122
0.309388
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
eae1bb293c1312b37011e03240675835981b85d2
121
py
Python
Hello world.py
KangFrank/Python_Start
5d09ef97dd3960d0cb1b99dca017fbc288dbc98c
[ "FSFAP" ]
2
2017-07-26T07:57:55.000Z
2018-09-30T14:39:21.000Z
Hello world.py
KangFrank/Python_Start
5d09ef97dd3960d0cb1b99dca017fbc288dbc98c
[ "FSFAP" ]
null
null
null
Hello world.py
KangFrank/Python_Start
5d09ef97dd3960d0cb1b99dca017fbc288dbc98c
[ "FSFAP" ]
null
null
null
#-*-coding:UTF-8-*- FileName:Hello world.py #python 2.6/2.7 print "Hello world.\n" #python 3.6 print("Hello world.\n")
13.444444
23
0.661157
23
121
3.478261
0.608696
0.375
0.375
0.4
0
0
0
0
0
0
0
0.065421
0.115702
121
8
24
15.125
0.682243
0.347107
0
0
0
0
0.368421
0
0
0
0
0
0
0
null
null
0
0
null
null
0.666667
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
1
0
7
d80f7ebbe37a732fe098e30ffb39ce28cc434721
3,201
py
Python
tests/test_lib.py
DK26/pyrust-keyring
f3a5aecf24c2ba7fbcf28e0d54eb0ef249c8fe1d
[ "Apache-2.0", "MIT" ]
1
2020-06-09T19:11:13.000Z
2020-06-09T19:11:13.000Z
tests/test_lib.py
DK26/pyrust-keyring
f3a5aecf24c2ba7fbcf28e0d54eb0ef249c8fe1d
[ "Apache-2.0", "MIT" ]
null
null
null
tests/test_lib.py
DK26/pyrust-keyring
f3a5aecf24c2ba7fbcf28e0d54eb0ef249c8fe1d
[ "Apache-2.0", "MIT" ]
null
null
null
import unittest from uuid import uuid4 import rskeyring def generate_password(): return str(uuid4()) class RSKeyringTests(unittest.TestCase): service_name = "rskeyring_unit_test_service_name" username = "rskeyring_unit_test_username" def test_set_password(self): rskeyring.set_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username, password=generate_password() ) def test_update_password(self): first_password = generate_password() rskeyring.set_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username, password=first_password ) stored_password = rskeyring.get_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username ) self.assertEqual(first_password, stored_password) second_password = generate_password() rskeyring.set_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username, password=second_password ) stored_password = rskeyring.get_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username ) self.assertEqual(second_password, stored_password) def test_get_password(self): generated_password = generate_password() rskeyring.set_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username, password=generated_password ) stored_password = rskeyring.get_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username ) self.assertEqual(stored_password, generated_password) def test_delete_password(self): generated_password = generate_password() rskeyring.set_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username, password=generated_password ) stored_password = rskeyring.get_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username ) self.assertEqual(stored_password, generated_password) rskeyring.delete_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username ) with self.assertRaises(OSError): rskeyring.get_password( service=RSKeyringTests.service_name, username=RSKeyringTests.username ) def test_exceptions(self): random_user_name = 'rskeyring_unittest_' + generate_password() with self.assertRaises(OSError): rskeyring.get_password( service=RSKeyringTests.service_name, username=random_user_name ) with self.assertRaises(OSError): rskeyring.delete_password( service=RSKeyringTests.service_name, username=random_user_name ) if __name__ == '__main__': unittest.main()
26.454545
70
0.646673
273
3,201
7.267399
0.135531
0.083165
0.134073
0.235887
0.777722
0.764113
0.764113
0.764113
0.71371
0.675403
0
0.00088
0.289597
3,201
120
71
26.675
0.871592
0
0
0.571429
1
0
0.027179
0.018744
0
0
0
0
0.083333
1
0.071429
false
0.380952
0.035714
0.011905
0.154762
0
0
0
0
null
0
0
1
0
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
0
0
1
0
0
0
0
0
8
d823e7061d051ef86b50d04b6b3e4bc08cba9ba2
73
py
Python
test/test.py
amraf1002/my-first-project
49d4af2e225b358ba7d2a4f97f78b8a208597ee7
[ "MIT" ]
null
null
null
test/test.py
amraf1002/my-first-project
49d4af2e225b358ba7d2a4f97f78b8a208597ee7
[ "MIT" ]
null
null
null
test/test.py
amraf1002/my-first-project
49d4af2e225b358ba7d2a4f97f78b8a208597ee7
[ "MIT" ]
null
null
null
from awesome_utility import hello print(hello()) print(hello('Frank'))
12.166667
33
0.753425
10
73
5.4
0.7
0.37037
0.555556
0
0
0
0
0
0
0
0
0
0.109589
73
5
34
14.6
0.830769
0
0
0
0
0
0.068493
0
0
0
0
0
0
1
0
true
0
0.333333
0
0.333333
0.666667
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
0
1
0
7
dc91c974ed98e20cdb5647d9665f79707049b4f9
141
py
Python
pandas_ml/core/__init__.py
matsavage/pandas-ml
794cddc8dc5d0a49fbc9734d826d9465078f376e
[ "BSD-3-Clause" ]
305
2016-02-21T06:35:25.000Z
2022-03-30T11:53:31.000Z
pandas_ml/core/__init__.py
matsavage/pandas-ml
794cddc8dc5d0a49fbc9734d826d9465078f376e
[ "BSD-3-Clause" ]
69
2016-02-16T08:10:46.000Z
2022-03-04T14:36:12.000Z
pandas_ml/core/__init__.py
matsavage/pandas-ml
794cddc8dc5d0a49fbc9734d826d9465078f376e
[ "BSD-3-Clause" ]
73
2016-02-16T08:27:28.000Z
2022-03-10T06:57:51.000Z
#!/usr/bin/env python from pandas_ml.core.frame import ModelFrame # noqa from pandas_ml.core.series import ModelSeries # noqa
28.2
57
0.716312
20
141
4.95
0.7
0.20202
0.242424
0.323232
0
0
0
0
0
0
0
0
0.205674
141
4
58
35.25
0.883929
0.212766
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
0
0
0
7
762db5e829492f610dce8dbc5dd54677af4b0814
56,407
py
Python
google/cloud/channel_v1/services/cloud_channel_service/pagers.py
renovate-bot/python-channel
0cebe664e7a5889269f63237f7c365656299e5d7
[ "Apache-2.0" ]
1
2021-11-13T12:15:04.000Z
2021-11-13T12:15:04.000Z
google/cloud/channel_v1/services/cloud_channel_service/pagers.py
renovate-bot/python-channel
0cebe664e7a5889269f63237f7c365656299e5d7
[ "Apache-2.0" ]
45
2021-01-14T23:02:15.000Z
2022-03-08T01:01:52.000Z
google/cloud/channel_v1/services/cloud_channel_service/pagers.py
renovate-bot/python-channel
0cebe664e7a5889269f63237f7c365656299e5d7
[ "Apache-2.0" ]
2
2021-01-14T22:39:56.000Z
2022-01-29T08:09:14.000Z
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from typing import ( Any, AsyncIterator, Awaitable, Callable, Sequence, Tuple, Optional, Iterator, ) from google.cloud.channel_v1.types import channel_partner_links from google.cloud.channel_v1.types import customers from google.cloud.channel_v1.types import entitlements from google.cloud.channel_v1.types import offers from google.cloud.channel_v1.types import products from google.cloud.channel_v1.types import service class ListCustomersPager: """A pager for iterating through ``list_customers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListCustomersResponse` object, and provides an ``__iter__`` method to iterate through its ``customers`` field. If there are more pages, the ``__iter__`` method will make additional ``ListCustomers`` requests and continue to iterate through the ``customers`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListCustomersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListCustomersResponse], request: service.ListCustomersRequest, response: service.ListCustomersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListCustomersRequest): The initial request object. response (google.cloud.channel_v1.types.ListCustomersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListCustomersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListCustomersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[customers.Customer]: for page in self.pages: yield from page.customers def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListCustomersAsyncPager: """A pager for iterating through ``list_customers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListCustomersResponse` object, and provides an ``__aiter__`` method to iterate through its ``customers`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListCustomers`` requests and continue to iterate through the ``customers`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListCustomersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListCustomersResponse]], request: service.ListCustomersRequest, response: service.ListCustomersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListCustomersRequest): The initial request object. response (google.cloud.channel_v1.types.ListCustomersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListCustomersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListCustomersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[customers.Customer]: async def async_generator(): async for page in self.pages: for response in page.customers: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListEntitlementsPager: """A pager for iterating through ``list_entitlements`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListEntitlementsResponse` object, and provides an ``__iter__`` method to iterate through its ``entitlements`` field. If there are more pages, the ``__iter__`` method will make additional ``ListEntitlements`` requests and continue to iterate through the ``entitlements`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListEntitlementsResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListEntitlementsResponse], request: service.ListEntitlementsRequest, response: service.ListEntitlementsResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListEntitlementsRequest): The initial request object. response (google.cloud.channel_v1.types.ListEntitlementsResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListEntitlementsRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListEntitlementsResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[entitlements.Entitlement]: for page in self.pages: yield from page.entitlements def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListEntitlementsAsyncPager: """A pager for iterating through ``list_entitlements`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListEntitlementsResponse` object, and provides an ``__aiter__`` method to iterate through its ``entitlements`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListEntitlements`` requests and continue to iterate through the ``entitlements`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListEntitlementsResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListEntitlementsResponse]], request: service.ListEntitlementsRequest, response: service.ListEntitlementsResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListEntitlementsRequest): The initial request object. response (google.cloud.channel_v1.types.ListEntitlementsResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListEntitlementsRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListEntitlementsResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[entitlements.Entitlement]: async def async_generator(): async for page in self.pages: for response in page.entitlements: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListTransferableSkusPager: """A pager for iterating through ``list_transferable_skus`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListTransferableSkusResponse` object, and provides an ``__iter__`` method to iterate through its ``transferable_skus`` field. If there are more pages, the ``__iter__`` method will make additional ``ListTransferableSkus`` requests and continue to iterate through the ``transferable_skus`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListTransferableSkusResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListTransferableSkusResponse], request: service.ListTransferableSkusRequest, response: service.ListTransferableSkusResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListTransferableSkusRequest): The initial request object. response (google.cloud.channel_v1.types.ListTransferableSkusResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListTransferableSkusRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListTransferableSkusResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[entitlements.TransferableSku]: for page in self.pages: yield from page.transferable_skus def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListTransferableSkusAsyncPager: """A pager for iterating through ``list_transferable_skus`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListTransferableSkusResponse` object, and provides an ``__aiter__`` method to iterate through its ``transferable_skus`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListTransferableSkus`` requests and continue to iterate through the ``transferable_skus`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListTransferableSkusResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListTransferableSkusResponse]], request: service.ListTransferableSkusRequest, response: service.ListTransferableSkusResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListTransferableSkusRequest): The initial request object. response (google.cloud.channel_v1.types.ListTransferableSkusResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListTransferableSkusRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListTransferableSkusResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[entitlements.TransferableSku]: async def async_generator(): async for page in self.pages: for response in page.transferable_skus: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListTransferableOffersPager: """A pager for iterating through ``list_transferable_offers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListTransferableOffersResponse` object, and provides an ``__iter__`` method to iterate through its ``transferable_offers`` field. If there are more pages, the ``__iter__`` method will make additional ``ListTransferableOffers`` requests and continue to iterate through the ``transferable_offers`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListTransferableOffersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListTransferableOffersResponse], request: service.ListTransferableOffersRequest, response: service.ListTransferableOffersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListTransferableOffersRequest): The initial request object. response (google.cloud.channel_v1.types.ListTransferableOffersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListTransferableOffersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListTransferableOffersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[service.TransferableOffer]: for page in self.pages: yield from page.transferable_offers def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListTransferableOffersAsyncPager: """A pager for iterating through ``list_transferable_offers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListTransferableOffersResponse` object, and provides an ``__aiter__`` method to iterate through its ``transferable_offers`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListTransferableOffers`` requests and continue to iterate through the ``transferable_offers`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListTransferableOffersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListTransferableOffersResponse]], request: service.ListTransferableOffersRequest, response: service.ListTransferableOffersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListTransferableOffersRequest): The initial request object. response (google.cloud.channel_v1.types.ListTransferableOffersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListTransferableOffersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListTransferableOffersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[service.TransferableOffer]: async def async_generator(): async for page in self.pages: for response in page.transferable_offers: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListChannelPartnerLinksPager: """A pager for iterating through ``list_channel_partner_links`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListChannelPartnerLinksResponse` object, and provides an ``__iter__`` method to iterate through its ``channel_partner_links`` field. If there are more pages, the ``__iter__`` method will make additional ``ListChannelPartnerLinks`` requests and continue to iterate through the ``channel_partner_links`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListChannelPartnerLinksResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListChannelPartnerLinksResponse], request: service.ListChannelPartnerLinksRequest, response: service.ListChannelPartnerLinksResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListChannelPartnerLinksRequest): The initial request object. response (google.cloud.channel_v1.types.ListChannelPartnerLinksResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListChannelPartnerLinksRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListChannelPartnerLinksResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[channel_partner_links.ChannelPartnerLink]: for page in self.pages: yield from page.channel_partner_links def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListChannelPartnerLinksAsyncPager: """A pager for iterating through ``list_channel_partner_links`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListChannelPartnerLinksResponse` object, and provides an ``__aiter__`` method to iterate through its ``channel_partner_links`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListChannelPartnerLinks`` requests and continue to iterate through the ``channel_partner_links`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListChannelPartnerLinksResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListChannelPartnerLinksResponse]], request: service.ListChannelPartnerLinksRequest, response: service.ListChannelPartnerLinksResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListChannelPartnerLinksRequest): The initial request object. response (google.cloud.channel_v1.types.ListChannelPartnerLinksResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListChannelPartnerLinksRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListChannelPartnerLinksResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[channel_partner_links.ChannelPartnerLink]: async def async_generator(): async for page in self.pages: for response in page.channel_partner_links: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListProductsPager: """A pager for iterating through ``list_products`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListProductsResponse` object, and provides an ``__iter__`` method to iterate through its ``products`` field. If there are more pages, the ``__iter__`` method will make additional ``ListProducts`` requests and continue to iterate through the ``products`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListProductsResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListProductsResponse], request: service.ListProductsRequest, response: service.ListProductsResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListProductsRequest): The initial request object. response (google.cloud.channel_v1.types.ListProductsResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListProductsRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListProductsResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[products.Product]: for page in self.pages: yield from page.products def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListProductsAsyncPager: """A pager for iterating through ``list_products`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListProductsResponse` object, and provides an ``__aiter__`` method to iterate through its ``products`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListProducts`` requests and continue to iterate through the ``products`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListProductsResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListProductsResponse]], request: service.ListProductsRequest, response: service.ListProductsResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListProductsRequest): The initial request object. response (google.cloud.channel_v1.types.ListProductsResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListProductsRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListProductsResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[products.Product]: async def async_generator(): async for page in self.pages: for response in page.products: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListSkusPager: """A pager for iterating through ``list_skus`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListSkusResponse` object, and provides an ``__iter__`` method to iterate through its ``skus`` field. If there are more pages, the ``__iter__`` method will make additional ``ListSkus`` requests and continue to iterate through the ``skus`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListSkusResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListSkusResponse], request: service.ListSkusRequest, response: service.ListSkusResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListSkusRequest): The initial request object. response (google.cloud.channel_v1.types.ListSkusResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListSkusRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListSkusResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[products.Sku]: for page in self.pages: yield from page.skus def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListSkusAsyncPager: """A pager for iterating through ``list_skus`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListSkusResponse` object, and provides an ``__aiter__`` method to iterate through its ``skus`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListSkus`` requests and continue to iterate through the ``skus`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListSkusResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListSkusResponse]], request: service.ListSkusRequest, response: service.ListSkusResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListSkusRequest): The initial request object. response (google.cloud.channel_v1.types.ListSkusResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListSkusRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListSkusResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[products.Sku]: async def async_generator(): async for page in self.pages: for response in page.skus: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListOffersPager: """A pager for iterating through ``list_offers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListOffersResponse` object, and provides an ``__iter__`` method to iterate through its ``offers`` field. If there are more pages, the ``__iter__`` method will make additional ``ListOffers`` requests and continue to iterate through the ``offers`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListOffersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListOffersResponse], request: service.ListOffersRequest, response: service.ListOffersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListOffersRequest): The initial request object. response (google.cloud.channel_v1.types.ListOffersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListOffersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListOffersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[offers.Offer]: for page in self.pages: yield from page.offers def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListOffersAsyncPager: """A pager for iterating through ``list_offers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListOffersResponse` object, and provides an ``__aiter__`` method to iterate through its ``offers`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListOffers`` requests and continue to iterate through the ``offers`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListOffersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListOffersResponse]], request: service.ListOffersRequest, response: service.ListOffersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListOffersRequest): The initial request object. response (google.cloud.channel_v1.types.ListOffersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListOffersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListOffersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[offers.Offer]: async def async_generator(): async for page in self.pages: for response in page.offers: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListPurchasableSkusPager: """A pager for iterating through ``list_purchasable_skus`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListPurchasableSkusResponse` object, and provides an ``__iter__`` method to iterate through its ``purchasable_skus`` field. If there are more pages, the ``__iter__`` method will make additional ``ListPurchasableSkus`` requests and continue to iterate through the ``purchasable_skus`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListPurchasableSkusResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListPurchasableSkusResponse], request: service.ListPurchasableSkusRequest, response: service.ListPurchasableSkusResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListPurchasableSkusRequest): The initial request object. response (google.cloud.channel_v1.types.ListPurchasableSkusResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListPurchasableSkusRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListPurchasableSkusResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[service.PurchasableSku]: for page in self.pages: yield from page.purchasable_skus def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListPurchasableSkusAsyncPager: """A pager for iterating through ``list_purchasable_skus`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListPurchasableSkusResponse` object, and provides an ``__aiter__`` method to iterate through its ``purchasable_skus`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListPurchasableSkus`` requests and continue to iterate through the ``purchasable_skus`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListPurchasableSkusResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListPurchasableSkusResponse]], request: service.ListPurchasableSkusRequest, response: service.ListPurchasableSkusResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListPurchasableSkusRequest): The initial request object. response (google.cloud.channel_v1.types.ListPurchasableSkusResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListPurchasableSkusRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListPurchasableSkusResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[service.PurchasableSku]: async def async_generator(): async for page in self.pages: for response in page.purchasable_skus: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListPurchasableOffersPager: """A pager for iterating through ``list_purchasable_offers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListPurchasableOffersResponse` object, and provides an ``__iter__`` method to iterate through its ``purchasable_offers`` field. If there are more pages, the ``__iter__`` method will make additional ``ListPurchasableOffers`` requests and continue to iterate through the ``purchasable_offers`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListPurchasableOffersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListPurchasableOffersResponse], request: service.ListPurchasableOffersRequest, response: service.ListPurchasableOffersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListPurchasableOffersRequest): The initial request object. response (google.cloud.channel_v1.types.ListPurchasableOffersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListPurchasableOffersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListPurchasableOffersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[service.PurchasableOffer]: for page in self.pages: yield from page.purchasable_offers def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListPurchasableOffersAsyncPager: """A pager for iterating through ``list_purchasable_offers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListPurchasableOffersResponse` object, and provides an ``__aiter__`` method to iterate through its ``purchasable_offers`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListPurchasableOffers`` requests and continue to iterate through the ``purchasable_offers`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListPurchasableOffersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListPurchasableOffersResponse]], request: service.ListPurchasableOffersRequest, response: service.ListPurchasableOffersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListPurchasableOffersRequest): The initial request object. response (google.cloud.channel_v1.types.ListPurchasableOffersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListPurchasableOffersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListPurchasableOffersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[service.PurchasableOffer]: async def async_generator(): async for page in self.pages: for response in page.purchasable_offers: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListSubscribersPager: """A pager for iterating through ``list_subscribers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListSubscribersResponse` object, and provides an ``__iter__`` method to iterate through its ``service_accounts`` field. If there are more pages, the ``__iter__`` method will make additional ``ListSubscribers`` requests and continue to iterate through the ``service_accounts`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListSubscribersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., service.ListSubscribersResponse], request: service.ListSubscribersRequest, response: service.ListSubscribersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiate the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListSubscribersRequest): The initial request object. response (google.cloud.channel_v1.types.ListSubscribersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListSubscribersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property def pages(self) -> Iterator[service.ListSubscribersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = self._method(self._request, metadata=self._metadata) yield self._response def __iter__(self) -> Iterator[str]: for page in self.pages: yield from page.service_accounts def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response) class ListSubscribersAsyncPager: """A pager for iterating through ``list_subscribers`` requests. This class thinly wraps an initial :class:`google.cloud.channel_v1.types.ListSubscribersResponse` object, and provides an ``__aiter__`` method to iterate through its ``service_accounts`` field. If there are more pages, the ``__aiter__`` method will make additional ``ListSubscribers`` requests and continue to iterate through the ``service_accounts`` field on the corresponding responses. All the usual :class:`google.cloud.channel_v1.types.ListSubscribersResponse` attributes are available on the pager. If multiple requests are made, only the most recent response is retained, and thus used for attribute lookup. """ def __init__( self, method: Callable[..., Awaitable[service.ListSubscribersResponse]], request: service.ListSubscribersRequest, response: service.ListSubscribersResponse, *, metadata: Sequence[Tuple[str, str]] = () ): """Instantiates the pager. Args: method (Callable): The method that was originally called, and which instantiated this pager. request (google.cloud.channel_v1.types.ListSubscribersRequest): The initial request object. response (google.cloud.channel_v1.types.ListSubscribersResponse): The initial response object. metadata (Sequence[Tuple[str, str]]): Strings which should be sent along with the request as metadata. """ self._method = method self._request = service.ListSubscribersRequest(request) self._response = response self._metadata = metadata def __getattr__(self, name: str) -> Any: return getattr(self._response, name) @property async def pages(self) -> AsyncIterator[service.ListSubscribersResponse]: yield self._response while self._response.next_page_token: self._request.page_token = self._response.next_page_token self._response = await self._method(self._request, metadata=self._metadata) yield self._response def __aiter__(self) -> AsyncIterator[str]: async def async_generator(): async for page in self.pages: for response in page.service_accounts: yield response return async_generator() def __repr__(self) -> str: return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
39.144344
88
0.670715
6,040
56,407
6.041556
0.039073
0.057877
0.046368
0.05152
0.957716
0.957716
0.957716
0.951961
0.946014
0.941355
0
0.003457
0.246086
56,407
1,440
89
39.171528
0.85463
0.455599
0
0.804044
0
0
0.007967
0
0
0
0
0
0
1
0.153966
false
0
0.010886
0.068429
0.284603
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
5201e84107f68ae07658aab43200f274c793872c
3,834
py
Python
forms.py
StarkerRegen/Pixelboooom
5b8239d66e9a9f4c890472f90d9fa9ffc593aa00
[ "MIT" ]
null
null
null
forms.py
StarkerRegen/Pixelboooom
5b8239d66e9a9f4c890472f90d9fa9ffc593aa00
[ "MIT" ]
3
2021-06-08T22:18:26.000Z
2022-03-12T00:46:44.000Z
forms.py
StarkerRegen/Pixelboooom
5b8239d66e9a9f4c890472f90d9fa9ffc593aa00
[ "MIT" ]
null
null
null
from flask_wtf import FlaskForm from wtforms import StringField, BooleanField, PasswordField, SubmitField, IntegerField, TextAreaField from wtforms.validators import DataRequired, Length, Email, Regexp, EqualTo, ValidationError from models import User class SigninForm(FlaskForm): # 域初始化时,第一个参数是设置label属性的 email = StringField('EmailAddress', validators=[DataRequired(), Email(message='Email address error.')]) password = PasswordField('Password', validators=[DataRequired(), Length(8,16), Regexp('^[a-zA-Z0-9]*$', message="The password should contain only a-z, A-z and 0-9." )]) remember_me = BooleanField('RememberMe', default=False) submit = SubmitField('Sign in') class SignupForm(FlaskForm): username = StringField('UserName', validators=[DataRequired(), Length(1,20), Regexp('^[a-zA-Z0-9]*$', message="The username should contain only a-z, A-z and 0-9." )]) email = StringField('EmailAddress', validators=[DataRequired(), Email(message='Email address error.')]) password = PasswordField('Password', validators=[DataRequired(), Length(8,16), Regexp('^[a-zA-Z0-9]*$', message="The password should contain only a-z, A-z and 0-9." )]) password_c = PasswordField('Password_c', validators=[DataRequired(), EqualTo('password',message='Password error')]) submit = SubmitField('Create account') def validate_username(self, username): user = User.query.filter_by(username=username.data).first() if user: raise ValidationError('User already exist.') def validate_email(self, email): user = User.query.filter_by(email=email.data).first() if user: raise ValidationError('Email has been registered, please re-enter') class ResetForm(FlaskForm): username = StringField('UserName', validators=[DataRequired(), Length(1,20), Regexp('^[a-zA-Z0-9]*$', message="The username should contain only a-z, A-z and 0-9." )]) email = StringField('EmailAddress', validators=[DataRequired(), Email(message='Email address error.')]) password = PasswordField('Password', validators=[DataRequired(), Length(8,16), Regexp('^[a-zA-Z0-9]*$', message="The password should contain only a-z, A-z and 0-9." )]) password_c = PasswordField('Password_c', validators=[DataRequired(), EqualTo('password',message='Password error')]) submit = SubmitField('Reset Passwd') def validate_username(self, username): user = User.query.filter_by(username=username.data).first() if not(user): raise ValidationError("User not registered.") def validate_email(self, email): user = User.query.filter_by(email=email.data).first() if not(user): raise ValidationError('User not registered.') class PostForm(FlaskForm): title = StringField('title', validators=[DataRequired(), Length(1,20,message='Length out of limit')]) category = StringField('category', validators=[DataRequired(), Length(1,50,message='Length out of limit')]) img = StringField('id', validators=[DataRequired(), Length(1,40,message='Length out of limit')]) style = StringField('style', validators=[DataRequired(), Length(1,10,message='Length out of limit')]) article = TextAreaField('style', validators=[DataRequired(), Length(1,120,message='Length out of limit')]) submit = SubmitField('Publish') class EditForm(FlaskForm): title = StringField('title', validators=[DataRequired(), Length(1,20,message='Length out of limit')]) category = StringField('category', validators=[DataRequired(), Length(1,50,message='Length out of limit')]) article = TextAreaField('style', validators=[DataRequired(), Length(1,120,message='Length out of limit')]) submit = SubmitField('Update')
58.984615
119
0.68362
447
3,834
5.832215
0.217002
0.151899
0.139624
0.111239
0.795167
0.773303
0.754891
0.754891
0.754891
0.754891
0
0.01908
0.166145
3,834
65
120
58.984615
0.796372
0.005738
0
0.611111
0
0
0.228811
0
0
0
0
0
0
1
0.074074
false
0.185185
0.074074
0
0.685185
0
0
0
0
null
0
0
0
0
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
1
0
0
1
0
0
7
5204fbad42cee1a49426d633df4f39a7e5239455
139
py
Python
app/monitor/__init__.py
bcarroll/inmoov_mini
029152cb186d45a040ddc71fd562191bff82b30c
[ "MIT" ]
null
null
null
app/monitor/__init__.py
bcarroll/inmoov_mini
029152cb186d45a040ddc71fd562191bff82b30c
[ "MIT" ]
null
null
null
app/monitor/__init__.py
bcarroll/inmoov_mini
029152cb186d45a040ddc71fd562191bff82b30c
[ "MIT" ]
null
null
null
from flask import Blueprint bp = Blueprint('monitor', __name__) from app.monitor import routes_memory from app.monitor import routes_cpu
19.857143
37
0.81295
20
139
5.35
0.55
0.130841
0.261682
0.373832
0.485981
0
0
0
0
0
0
0
0.129496
139
6
38
23.166667
0.884298
0
0
0
0
0
0.05036
0
0
0
0
0
0
1
0
false
0
0.75
0
0.75
0.5
1
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
1
0
1
1
0
7
52273f979806993100f37ec772519c92c772855f
1,799
py
Python
src/models.py
ryzenboi98/btc-data-api
e9d7447595a35834bf63e738e797fb8b26c9a134
[ "MIT" ]
null
null
null
src/models.py
ryzenboi98/btc-data-api
e9d7447595a35834bf63e738e797fb8b26c9a134
[ "MIT" ]
null
null
null
src/models.py
ryzenboi98/btc-data-api
e9d7447595a35834bf63e738e797fb8b26c9a134
[ "MIT" ]
null
null
null
from sqlalchemy import Integer, String, Float, DateTime from sqlalchemy.sql.expression import null from sqlalchemy.sql.schema import Column from src.database import Base class Stats(Base): __tablename__ = 'stats' id = Column(Integer, primary_key=True) lower = Column(Float, nullable=False) higher = Column(Float, nullable=False) open = Column(Float, nullable=False) close = Column(Float, nullable=False) volume = Column(Float, nullable=False) change = Column(Float, nullable=True) timestamp = Column(DateTime, nullable=False) class WeeklyStats(Base): __tablename__ = 'weekly_stats' id = Column(Integer, primary_key=True) lower = Column(Float, nullable=False) higher = Column(Float, nullable=False) open = Column(Float, nullable=False) close = Column(Float, nullable=False) avg_volume = Column(Float, nullable=False) change = Column(Float, nullable=True) timestamp = Column(DateTime, nullable=False) class MonthlyStats(Base): __tablename__ = 'monthly_stats' id = Column(Integer, primary_key=True) lower = Column(Float, nullable=False) higher = Column(Float, nullable=False) open = Column(Float, nullable=False) close = Column(Float, nullable=False) avg_volume = Column(Float, nullable=False) change = Column(Float, nullable=True) timestamp = Column(DateTime, nullable=False) class AnualStats(Base): __tablename__ = 'anual_stats' id = Column(Integer, primary_key=True) lower = Column(Float, nullable=False) higher = Column(Float, nullable=False) open = Column(Float, nullable=False) close = Column(Float, nullable=False) avg_volume = Column(Float, nullable=False) change = Column(Float, nullable=True) timestamp = Column(DateTime, nullable=False)
32.709091
55
0.714842
214
1,799
5.88785
0.17757
0.209524
0.361905
0.380952
0.796825
0.796825
0.796825
0.796825
0.796825
0.796825
0
0
0.178432
1,799
54
56
33.314815
0.852503
0
0
0.704545
0
0
0.022803
0
0
0
0
0
0
1
0
false
0
0.090909
0
1
0
0
0
0
null
1
1
1
0
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
9
523adcb3b588cbbb6467966415118284722b39b4
3,641
py
Python
snipeit/Fields.py
adfedotov/snipeit-api
0f74a4a1957f6e1bd247af714309d429d38c7331
[ "MIT" ]
null
null
null
snipeit/Fields.py
adfedotov/snipeit-api
0f74a4a1957f6e1bd247af714309d429d38c7331
[ "MIT" ]
null
null
null
snipeit/Fields.py
adfedotov/snipeit-api
0f74a4a1957f6e1bd247af714309d429d38c7331
[ "MIT" ]
null
null
null
import requests as r class Fields(): def __init__(self, server, headers): self.server = server self.headers = headers def get(self): """Get custom fields Returns: dict: Json response """ endpoint = self.server + '/api/v1/fields' response = r.request('GET', endpoint, headers=self.headers) return response.json() def get_field_by_id(self, id: int): """Get field by id Args: id (int): field id Returns: dict: Json response """ endpoint = f'{self.server}/api/v1/fields/{id}' response = r.request('GET', endpoint, headers=self.headers) return response.json() def create(self, payload: dict): """Create field Args: payload (dict): Payload has to contain 'name', 'element'. Accepted: # name (str) # element (str) [text, textarea, checkbox, radio, listbox] # field_values (str) # show_in_email (bool) # format (str) # field_encrypted (bool) # help_text (str) Returns: dict: Json response """ endpoint = f'{self.server}/api/v1/fields' response = r.request('POST', endpoint, headers=self.headers, json=payload) return response.json() def update(self, id: int, payload: dict): """Update field Args: id (int): field ID Payload has to contain 'name', 'element'. Accepted: # name (str) # element (str) [text, textarea, checkbox, radio, listbox] # field_values (str) # show_in_email (bool) # format (str) # field_encrypted (bool) # help_text (str) Returns: dict: Json response """ endpoint = f'{self.server}/api/v1/fields/{id}' response = r.request('PUT', endpoint, headers=self.headers, json=payload) return response.json() def associate_with(self, id: int, fieldset_id: int): """Associate custom field with a custom fieldset Args: id (int): field id fieldset_id (int): fieldset id Returns: dict: Json response """ payload = { 'fieldset_id': fieldset_id } endpoint = f'{self.server}/api/v1/fields/{id}/associate' response = r.request('POST', endpoint, headers=self.headers, json=payload) return response.json() def disassociate_with(self, id: int, fieldset_id: int): """Disassociate custom field with a custom fieldset Args: id (int): field id fieldset_id (int): fieldset id Returns: dict: Json response """ payload = { 'fieldset_id': fieldset_id } endpoint = f'{self.server}/api/v1/fields/{id}/disassociate' response = r.request('POST', endpoint, headers=self.headers, json=payload) return response.json() def delete(self, id: int): """Delete field Args: id (int): field id Returns: dict: Json response """ endpoint = f'{self.server}/api/v1/fields/{id}' response = r.request('DELETE', endpoint, headers=self.headers) return response.json()
27.171642
86
0.503433
368
3,641
4.907609
0.160326
0.03876
0.05814
0.089147
0.825028
0.807863
0.793466
0.74031
0.728682
0.728682
0
0.003142
0.38808
3,641
134
87
27.171642
0.807451
0.381214
0
0.487179
0
0
0.150745
0.115958
0
0
0
0
0
1
0.205128
false
0
0.025641
0
0.435897
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
0
0
0
0
0
7
524b02c3b4cfe186f11492b2be2e296ebfec1a32
22,094
py
Python
tests/integration/views/surveys/help/test_surveys_help_info_about_this_survey.py
ONSdigital/ras-frontstage
e8ff1931b49cb3ab47b421aed6780e9e944dceea
[ "MIT" ]
8
2017-06-30T12:32:02.000Z
2022-02-25T09:07:28.000Z
tests/integration/views/surveys/help/test_surveys_help_info_about_this_survey.py
ONSdigital/ras-frontstage
e8ff1931b49cb3ab47b421aed6780e9e944dceea
[ "MIT" ]
256
2017-05-16T09:38:09.000Z
2022-03-28T13:38:42.000Z
tests/integration/views/surveys/help/test_surveys_help_info_about_this_survey.py
ONSdigital/ras-frontstage
e8ff1931b49cb3ab47b421aed6780e9e944dceea
[ "MIT" ]
4
2017-09-29T08:58:36.000Z
2021-04-11T07:44:27.000Z
import unittest from unittest.mock import patch import requests_mock from frontstage import app from tests.integration.mocked_services import ( business_party, encoded_jwt_token, respondent_party, survey, survey_eq, survey_list_todo, survey_rsi, url_banner_api, ) class TestSurveyHelpInfoAboutThisSurvey(unittest.TestCase): def setUp(self): self.app = app.test_client() self.app.set_cookie("localhost", "authorization", "session_key") self.headers = { "Authorization": "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJ1c2VyX2lkIjoicmluZ3JhbUBub3d3aGVyZS5jb20iLCJ1c2Vy" "X3Njb3BlcyI6WyJjaS5yZWFkIiwiY2kud3JpdGUiXX0.se0BJtNksVtk14aqjp7SvnXzRbEKoqXb8Q5U9VVdy54" # NOQA } self.patcher = patch("redis.StrictRedis.get", return_value=encoded_jwt_token) self.patcher.start() def tearDown(self): self.patcher.stop() def set_flask_session(self): with self.app.session_transaction() as mock_session: mock_session["help_survey_ref"] = "074" mock_session["help_ru_ref"] = "49900000001F" @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_qbs(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey_eq get_business.return_value = business_party response = self.app.get("/surveys/surveys-help?survey_ref=139&ru_ref=49900000001F", follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Help".encode(), response.data) self.assertIn("Choose an option".encode(), response.data) self.assertIn("Information about the Quarterly Business Survey".encode(), response.data) self.assertIn("Continue".encode(), response.data) self.assertIn("Cancel".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_bricks(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party response = self.app.get("/surveys/surveys-help?survey_ref=074&ru_ref=49900000001F", follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Help".encode(), response.data) self.assertIn("Choose an option".encode(), response.data) self.assertIn("Information about the Monthly Survey of Building Materials Bricks".encode(), response.data) self.assertIn("Continue".encode(), response.data) self.assertIn("Cancel".encode(), response.data) self.assertIn("Help".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_bricks_with_no_option_select(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party form = {} self.set_flask_session() response = self.app.post("/surveys/help", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("There is 1 error on this page".encode(), response.data) self.assertIn("You need to choose an option".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_bricks_with_option_select(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party form = {"option": "info-about-this-survey"} self.set_flask_session() response = self.app.post("/surveys/help", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Information about the Monthly Survey of Building Materials Bricks".encode(), response.data) self.assertIn("Choose an option".encode(), response.data) self.assertIn("Can I be exempt from completing the survey questionnaire?".encode(), response.data) self.assertIn("How was my business selected?".encode(), response.data) self.assertIn("How long will it take to complete?".encode(), response.data) self.assertIn("How long will my business be selected for?".encode(), response.data) self.assertIn("Something else".encode(), response.data) self.assertIn("Continue".encode(), response.data) self.assertIn("Cancel".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_bricks_with_sub_option_exemption_completing_survey( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party form = {"option": "exemption-completing-survey"} self.set_flask_session() response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Can I be exempt from completing the survey questionnaire?".encode(), response.data) self.assertIn( "While this survey is voluntary, we have selected your company in the same way we ".encode(), response.data ) self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data) self.assertIn("Did this answer your question?".encode(), response.data) self.assertIn("Yes".encode(), response.data) self.assertIn("No".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_rsi_with_sub_option_exemption_completing_survey( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey_rsi get_business.return_value = business_party form = {"option": "exemption-completing-survey"} self.set_flask_session() response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Can I be exempt from completing the survey questionnaire?".encode(), response.data) self.assertIn( "No. Once selected, the law obliges a business to complete the survey under the provisions " "of the Statistics of Trade Act 1947".encode(), response.data, ) self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data) self.assertIn("Did this answer your question?".encode(), response.data) self.assertIn("Yes".encode(), response.data) self.assertIn("No".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_bricks_with_sub_option_why_selected(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party form = {"option": "why-selected"} self.set_flask_session() response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("How was my business selected?".encode(), response.data) self.assertIn( "We select businesses from the Inter-Departmental Business Register (IDBR). ".encode(), response.data ) self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data) self.assertIn("Did this answer your question?".encode(), response.data) self.assertIn("Yes".encode(), response.data) self.assertIn("No".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_bricks_with_sub_option_time_to_complete(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party form = {"option": "time-to-complete"} self.set_flask_session() response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("How long will it take to complete?".encode(), response.data) self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data) self.assertIn("Did this answer your question?".encode(), response.data) self.assertIn("Yes".encode(), response.data) self.assertIn("No".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_bricks_with_sub_option_how_long_selected_for( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party form = {"option": "how-long-selected-for"} self.set_flask_session() response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("How long will my business be selected for?".encode(), response.data) self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data) self.assertIn("Did this answer your question?".encode(), response.data) self.assertIn("Yes".encode(), response.data) self.assertIn("No".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_for_voluntary_survey_with_sub_option_penalties(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party form = {"option": "penalties"} self.set_flask_session() response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Are there penalties for not completing this survey?".encode(), response.data) self.assertIn("Did this answer your question?".encode(), response.data) self.assertNotIn( "If you do not contact us or complete and return by the deadline, penalties may be incurred " "resulting in a fine of up to £2,500 (under section 4 of the Statistics of Trade Act 1947, " "last updated by section 17 of the Criminal Justice Act 1991).".encode(), response.data, ) self.assertIn("Yes".encode(), response.data) self.assertIn("No".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_for_statutory_survey_with_sub_option_penalties(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey_eq get_business.return_value = business_party form = {"option": "penalties"} self.set_flask_session() response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Are there penalties for not completing this survey?".encode(), response.data) self.assertIn("Did this answer your question?".encode(), response.data) self.assertIn( "If you do not contact us or complete and return by the deadline, penalties may be incurred " "resulting in a fine of up to £2,500 (under section 4 of the Statistics of Trade Act 1947, " "last updated by section 17 of the Criminal Justice Act 1991).".encode(), response.data, ) self.assertIn("Yes".encode(), response.data) self.assertIn("No".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_info_bricks_with_sub_option_something_else(self, mock_request, get_survey, get_business): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party form = {"option": "info-something-else"} self.set_flask_session() response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Information about the Monthly Survey of Building Materials Bricks".encode(), response.data) self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data) self.assertIn("Did this answer your question?".encode(), response.data) self.assertIn("Yes".encode(), response.data) self.assertIn("No".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_send_message_info_bricks_with_sub_option_exemption_completing_survey( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party self.set_flask_session() response = self.app.get( "/surveys/help/info-about-this-survey/" "exemption-completing-survey/send-message", follow_redirects=True, ) self.assertEqual(response.status_code, 200) self.assertIn("Can I be exempt from completing the survey questionnaire?".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_send_message_info_bricks_with_sub_option_why_selected_survey( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party self.set_flask_session() response = self.app.get( "/surveys/help/info-about-this-survey/" "why-selected/send-message", follow_redirects=True ) self.assertEqual(response.status_code, 200) self.assertIn("How was my business selected?".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_send_message_info_bricks_with_sub_option_time_to_complete( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party self.set_flask_session() response = self.app.get( "/surveys/help/info-about-this-survey/" "time-to-complete/send-message", follow_redirects=True, ) self.assertEqual(response.status_code, 200) self.assertIn("How long will it take to complete?".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_send_message_info_bricks_with_sub_option_how_long_selected_for( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party self.set_flask_session() response = self.app.get( "/surveys/help/info-about-this-survey/how-long-selected-for/send-message", follow_redirects=True, ) self.assertEqual(response.status_code, 200) self.assertIn("Send a message".encode(), response.data) self.assertIn("Describe your issue and we will get back to you.".encode(), response.data) self.assertIn("How long will my business be selected for?".encode(), response.data) self.assertIn("Enter message".encode(), response.data) self.assertIn("Send message".encode(), response.data) self.assertIn("Cancel".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_send_message_info_bricks_with_sub_option_penalties( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party self.set_flask_session() response = self.app.get("/surveys/help/info-about-this-survey/penalties/send-message", follow_redirects=True) self.assertEqual(response.status_code, 200) self.assertIn("Send a message".encode(), response.data) self.assertIn("Describe your issue and we will get back to you.".encode(), response.data) self.assertIn("What are the penalties for not completing a survey?".encode(), response.data) self.assertIn("Enter message".encode(), response.data) self.assertIn("Send message".encode(), response.data) self.assertIn("Cancel".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_survey_help_send_message_info_bricks_with_sub_option_info_something_else( self, mock_request, get_survey, get_business ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party self.set_flask_session() response = self.app.get( "/surveys/help/info-about-this-survey/info-something-else/send-message", follow_redirects=True, ) self.assertEqual(response.status_code, 200) self.assertIn("Send a message".encode(), response.data) self.assertIn("Describe your issue and we will get back to you.".encode(), response.data) self.assertIn("Information about this survey".encode(), response.data) self.assertIn("Enter message".encode(), response.data) self.assertIn("Send message".encode(), response.data) self.assertIn("Cancel".encode(), response.data) @requests_mock.mock() @patch("frontstage.controllers.party_controller.get_respondent_party_by_id") @patch("frontstage.controllers.party_controller.get_survey_list_details_for_party") @patch("frontstage.controllers.conversation_controller.send_message") @patch("frontstage.controllers.party_controller.get_business_by_ru_ref") @patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref") def test_create_message_post_success( self, mock_request, get_survey, get_business, send_message, get_survey_list, get_respondent_party_by_id ): mock_request.get(url_banner_api, status_code=404) get_survey.return_value = survey get_business.return_value = business_party get_survey_list.return_value = survey_list_todo get_respondent_party_by_id.return_value = respondent_party form = {"body": "info-something-else"} self.set_flask_session() response = self.app.post( "/surveys/help/info-about-this-survey/info-something-else/send-message", data=form, follow_redirects=True, ) self.assertEqual(response.status_code, 200) self.assertIn("Message sent.".encode(), response.data)
52.47981
120
0.717435
2,785
22,094
5.42298
0.075404
0.080646
0.103688
0.099053
0.90386
0.896511
0.889889
0.882871
0.882871
0.882871
0
0.01219
0.175704
22,094
420
121
52.604762
0.816989
0.000181
0
0.748011
0
0.013263
0.309716
0.171043
0
0
0
0
0.281167
1
0.058355
false
0
0.013263
0
0.074271
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
bff941c9a17ec2e4cc0657fbf59116362103afa3
6,250
py
Python
config/sparql_queries.py
ficolo/science-radar
3929493d93aa3712e4f48e7ab23b659daedb3b1f
[ "Apache-2.0" ]
1
2019-10-03T02:04:12.000Z
2019-10-03T02:04:12.000Z
config/sparql_queries.py
ficolo/science-radar
3929493d93aa3712e4f48e7ab23b659daedb3b1f
[ "Apache-2.0" ]
null
null
null
config/sparql_queries.py
ficolo/science-radar
3929493d93aa3712e4f48e7ab23b659daedb3b1f
[ "Apache-2.0" ]
null
null
null
queries = { 'CO-AUTHORSHIP': """ PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> PREFIX foaf: <http://xmlns.com/foaf/0.1/> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX bibo: <http://purl.org/ontology/bibo/> PREFIX sio: <http://semanticscience.org/resource/> SELECT ?node1 ?node2 ?date (COUNT (DISTINCT ?paper) as ?weight){{ {{ SELECT ?paper ?title ?date {{ ?paper sio:SIO_001278 ?dataset . ?paper dcterms:title ?title . ?paper dcterms:issued ?date . FILTER(?date >= "{start_year}-{start_month}-01T00:00:00"^^xsd:dateTime && ?date < "{end_year}-{end_month}-01T00:00:00"^^xsd:dateTime) }} }} ?paper bibo:authorList ?authorList . ?authorList rdfs:member ?author1 . ?authorList rdfs:member ?author2 . ?author1 foaf:name ?node1 . ?author2 foaf:name ?node2 . FILTER(?author1 != ?author2) }} GROUP BY ?node1 ?node2 ?date """, 'BURST': """ PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX sio: <http://semanticscience.org/resource/> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX oa: <http://www.w3.org/ns/oa#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX owl: <http://www.w3.org/2002/07/owl#> PREFIX doco: <http://purl.org/spar/doco/> PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX bibo: <http://purl.org/ontology/bibo/> SELECT ?date ?title ?abstract (GROUP_CONCAT(DISTINCT ?annotation1Label;separator="|") AS ?annotations) {{ ?paper sio:SIO_001278 ?dataset . ?paper dcterms:title ?title . ?paper bibo:abstract ?abstract . ?paper dcterms:issued ?date . ?annotation1 a oa:Annotation ; oa:hasTarget ?paragraph ; oa:hasBody ?ontoBody1 ; oa:hasBody ?textualBody1 . ?textualBody1 a oa:TextualBody ; rdf:value ?annotation1Label . ?paragraph oa:hasSource ?paper . FILTER(?date >= "{start}-01-01T00:00:00"^^xsd:dateTime && ?date < "{end}-01-01T00:00:00"^^xsd:dateTime) FILTER(STRSTARTS(STR(?ontoBody1), "{ontology}")) }} GROUP BY ?date ?title ?abstract """, 'PAPER_ANNOTATIONS': """ PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX sio: <http://semanticscience.org/resource/> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX oa: <http://www.w3.org/ns/oa#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX luc: <http://www.ontotext.com/owlim/lucene#> PREFIX owl: <http://www.w3.org/2002/07/owl#> PREFIX doco: <http://purl.org/spar/doco/> PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX bibo: <http://purl.org/ontology/bibo/> SELECT ?date ?title (GROUP_CONCAT(DISTINCT ?annotation1Label;separator="|") AS ?annotations) {{ {{ SELECT ?paper ?title ?date {{ ?paper sio:SIO_001278 ?dataset . ?paper dcterms:title ?title . ?paper dcterms:issued ?date . FILTER(?date >= "{start_year}-{start_month}-01T00:00:00"^^xsd:dateTime && ?date < "{end_year}-{end_month}-01T00:00:00"^^xsd:dateTime) }} }} ?paragraph oa:hasSource ?paper . ?annotation1 a oa:Annotation ; oa:hasTarget ?paragraph ; oa:hasBody ?ontoBody1 ; oa:hasBody ?textualBody1 . ?textualBody1 a oa:TextualBody ; rdf:value ?annotation1Label . FILTER(STRSTARTS(STR(?ontoBody1), "http://ncicb.nci.nih.gov/xml/owl/EVS/Thesaurus.owl")) }} GROUP BY ?date ?title ?abstract """, 'CO-CITATION': """ PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX sio: <http://semanticscience.org/resource/> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX oa: <http://www.w3.org/ns/oa#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX luc: <http://www.ontotext.com/owlim/lucene#> PREFIX owl: <http://www.w3.org/2002/07/owl#> PREFIX doco: <http://purl.org/spar/doco/> PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX bibo: <http://purl.org/ontology/bibo/> SELECT ?title1 ?title2 (COUNT (DISTINCT ?paper) as ?weight) { { SELECT ?paper { ?paper sio:SIO_001278 ?dataset . ?paper dcterms:issued ?date . FILTER(?date >= "{start}-01-01T00:00:00"^^xsd:dateTime && ?date < "{end}-01-01T00:00:00"^^xsd:dateTime) } } ?paper bibo:cites ?citedPaper1 . ?paper bibo:cites ?citedPaper2 . ?citedPaper1 dcterms:title ?title1 . ?citedPaper2 dcterms:title ?title2 . } GROUP BY ?title1 ?title2 """, 'PAPER-CITATIONS': """ PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX sio: <http://semanticscience.org/resource/> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX oa: <http://www.w3.org/ns/oa#> PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX luc: <http://www.ontotext.com/owlim/lucene#> PREFIX owl: <http://www.w3.org/2002/07/owl#> PREFIX doco: <http://purl.org/spar/doco/> PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX bibo: <http://purl.org/ontology/bibo/> SELECT ?date ?title (GROUP_CONCAT (DISTINCT ?citedPaper; separator="|") as ?references) {{ {{ SELECT ?date ?paper ?title {{ ?paper sio:SIO_001278 ?dataset . ?paper dcterms:issued ?date . ?paper dcterms:title ?title . FILTER(?date >= "{start_year}-{start_month}-01T00:00:00"^^xsd:dateTime && ?date < "{end_year}-{end_month}-01T00:00:00"^^xsd:dateTime) }} }} ?paper bibo:cites ?citedPaper . ?citedPaper a bibo:AcademicArticle . }} GROUP BY ?date ?title """ }
44.326241
149
0.60064
797
6,250
4.683814
0.136763
0.05813
0.067506
0.090008
0.798286
0.785427
0.77498
0.725422
0.712028
0.676668
0
0.066982
0.22128
6,250
141
150
44.326241
0.700021
0
0
0.722628
0
0.116788
0.984962
0.102224
0
0
0
0
0
1
0
false
0
0
0
0
0
0
0
0
null
0
0
0
0
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
872ab1779654eb50d33ff0c304ef10703b774223
19,028
py
Python
bayes_layers.py
JurijsNazarovs/bayesian_nn
936bf55e0a1e620504d5159c100a74493bd16399
[ "MIT" ]
6
2021-08-25T04:26:41.000Z
2022-01-24T01:29:40.000Z
bayes_layers.py
JurijsNazarovs/bayesian_nn
936bf55e0a1e620504d5159c100a74493bd16399
[ "MIT" ]
1
2021-10-09T07:29:21.000Z
2021-10-11T16:59:38.000Z
bayes_layers.py
JurijsNazarovs/bayesian_nn
936bf55e0a1e620504d5159c100a74493bd16399
[ "MIT" ]
3
2021-09-05T08:58:20.000Z
2021-12-03T14:54:09.000Z
import torch from torch import nn from torch.nn import Parameter import torch.nn.functional as F from torch.nn.modules.utils import _single, _pair, _triple import numpy as np import math import vi_posteriors as vip import importlib importlib.reload(vip) minvar = -3 maxvar = -2 # ------------------------------------------------------------------------------- # Main building blocks # ------------------------------------------------------------------------------- class BayesLocScaleConvBaseBlock(nn.Module): """ Necessary to define this block, despite that it looks like _ConvNd from pyTorch, because we are training posterior parameters to sample weights and not weights by itself. Thus, self.weights are not presented here. We assume that approximate posterior of weights can be parametrized by 2 parameters: mu and logsigmasq. """ def __init__(self, in_channels, out_channels, kernel_size, stride, padding, dilation, transposed, groups, prior_mu=0, prior_logsigmasq=0, is_mixed_prior=False, prior_p_mixed=1 / 2, bias=False): # Default description of module _ConvNd from pytorch super().__init__() if in_channels % groups != 0: raise ValueError('in_channels must be divisible by groups') if out_channels % groups != 0: raise ValueError('out_channels must be divisible by groups') self.in_channels = in_channels self.out_channels = out_channels self.kernel_size = kernel_size self.stride = stride self.padding = padding self.dilation = dilation self.groups = groups self.bias = bias # Save parameters to fill prior self.prior_params = {} self.prior_params['mu'] = prior_mu self.prior_params['logsigmasq'] = prior_logsigmasq self.prior_params['is_mixed'] = is_mixed_prior self.prior_params['p_mixed'] = prior_p_mixed # Posterior parameters - trained: mu and logsigmasq if transposed: self.post_mu = Parameter( torch.zeros( (in_channels, out_channels // groups, *kernel_size))) self.post_logsigmasq = Parameter( torch.zeros( (in_channels, out_channels // groups, *kernel_size))) else: self.post_mu = Parameter( torch.zeros( (out_channels, in_channels // groups, *kernel_size))) self.post_logsigmasq = Parameter( torch.zeros( (out_channels, in_channels // groups, *kernel_size))) if self.bias: self.post_mu_bias = Parameter( torch.zeros((out_channels, ) + (1, ) * len(kernel_size))) self.post_logsigmasq_bias = Parameter( torch.zeros((out_channels, ) + (1, ) * len(kernel_size))) # Initialize parameters self.reset_parameters() self.reset_priors() def reset_parameters(self): n = self.in_channels for k in self.kernel_size: n *= k stdv = 1. / math.sqrt(n) self.post_mu.data.uniform_(-stdv, stdv) self.post_logsigmasq.data.uniform_(minvar, maxvar) # (0, 1) if self.bias: self.post_mu_bias.data.uniform_(-stdv, stdv) self.post_logsigmasq_bias.data.uniform_(minvar, maxvar) # (0, 1) def reset_priors(self): if self.prior_params['is_mixed']: # Scale mixture prior logsigmasq1 = -2 logsigmasq2 = -8 u = torch.rand(self.post_logsigmasq.shape, dtype=self.post_logsigmasq.dtype, requires_grad=False).to("cpu") mod_ind = torch.tensor(u <= self.prior_params['p_mixed'], dtype=self.post_logsigmasq.dtype) self.prior_mu = Parameter(torch.zeros_like(self.post_mu) + self.prior_params['mu'], requires_grad=False) self.prior_logsigmasq = Parameter(mod_ind * logsigmasq1 + (1 - mod_ind) * logsigmasq2, requires_grad=False) if self.bias: self.prior_mu_bias = Parameter( torch.zeros_like(self.post_mu_bias) + self.prior_params['mu'], requires_grad=False) u = torch.rand(self.post_logsigmasq_bias.shape, dtype=self.post_logsigmasq.dtype).to("cpu") mod_ind = torch.tensor(u <= self.prior_params['p_mixed'], dtype=self.post_logsigmasq.dtype) self.prior_logsigmasq_bias = Parameter( mod_ind * logsigmasq1 + (1 - mod_ind) * logsigmasq2, requires_grad=False) else: self.prior_mu = Parameter(torch.zeros_like(self.post_mu) + self.prior_params['mu'], requires_grad=False) self.prior_logsigmasq = Parameter( torch.zeros_like(self.post_logsigmasq) + self.prior_params['logsigmasq'], requires_grad=False) if self.bias: self.prior_mu_bias = Parameter( torch.zeros_like(self.post_mu_bias) + self.prior_params['mu'], requires_grad=False) self.prior_logsigmasq_bias = Parameter( torch.zeros_like(self.post_logsigmasq_bias) + self.prior_params['logsigmasq'], requires_grad=False) def extra_repr(self): # Displays arguments for a block s = ('{in_channels}, {out_channels}, kernel_size={kernel_size}, ' 'stride={stride}') if self.padding != (0, ) * len(self.padding): s += ', padding={padding}' if self.dilation != (1, ) * len(self.dilation): s += ', dilation={dilation}' if self.groups != 1: s += ', groups={groups}' s += ', bias = {bias}' if hasattr(self, "activation"): s += ', activation={activation}' if hasattr(self, "approx_post"): s += ', approx_post={approx_post}' if hasattr(self, "kl_method"): s += ', kl_method={kl_method}' if hasattr(self, "n_mc_iter"): s += ', n_mc_iter={n_mc_iter}' if hasattr(self, "compute_kl"): s += ', compute_kl={compute_kl}' return s.format(**self.__dict__) class Conv1d(BayesLocScaleConvBaseBlock): """ Bayesian Convolution 1d layer. Learned parameters are mu and logsigmasq. """ def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, dilation=1, activation=None, is_mixed_prior=False, bias=False, approx_post="Radial", kl_method="repar", n_mc_iter=20, **kwargs): kernel_size = _single(kernel_size) stride = _single(stride) padding = _single(padding) dilation = _single(dilation) self.activation = activation self.approx_post = approx_post self.kl_method = kl_method self.n_mc_iter = n_mc_iter self.compute_kl = True super().__init__(in_channels, out_channels, kernel_size, stride, padding, dilation, False, groups, is_mixed_prior=is_mixed_prior, bias=bias) def forward(self, x): post = eval("vip." + self.approx_post) output, kl = post.forward(self, x, fun="conv1d") return output, kl class Conv2d(BayesLocScaleConvBaseBlock): """ Bayesian Convolution 2d layer. Learned parameters are mu and logsigmasq. """ def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, dilation=1, activation=None, is_mixed_prior=False, bias=False, approx_post="Radial", kl_method="repar", n_mc_iter=20, **kwargs): kernel_size = _pair(kernel_size) stride = _pair(stride) padding = _pair(padding) dilation = _pair(dilation) self.activation = activation self.approx_post = approx_post self.kl_method = kl_method self.n_mc_iter = n_mc_iter self.compute_kl = True super().__init__(in_channels, out_channels, kernel_size, stride, padding, dilation, False, groups, is_mixed_prior=is_mixed_prior, bias=bias) def forward(self, x): post = eval("vip." + self.approx_post) output, kl = post.forward(self, x, fun="conv2d") return output, kl class Conv3d(BayesLocScaleConvBaseBlock): """ Bayesian Convolution 3d layer. Learned parameters are mu and logsigmasq. """ def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, dilation=1, activation=None, is_mixed_prior=False, bias=False, approx_post="Radial", kl_method="repar", n_mc_iter=20, **kwargs): kernel_size = _triple(kernel_size) stride = _triple(stride) padding = _triple(padding) dilation = _triple(dilation) self.activation = activation self.approx_post = approx_post self.kl_method = kl_method self.n_mc_iter = n_mc_iter self.compute_kl = True super().__init__(in_channels, out_channels, kernel_size, stride, padding, dilation, False, groups, is_mixed_prior=is_mixed_prior, bias=bias) def forward(self, x): post = eval("vip." + self.approx_post) output, kl = post.forward(self, x, fun="conv3d") return output, kl # Transpose classes class ConvTranspose1d(BayesLocScaleConvBaseBlock): """ Bayesian Convolution 1d layer. Learned parameters are mu and logsigmasq. """ def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, dilation=1, activation=None, is_mixed_prior=False, bias=False, approx_post="Radial", kl_method="repar", n_mc_iter=20, **kwargs): kernel_size = _single(kernel_size) stride = _single(stride) padding = _single(padding) dilation = _single(dilation) self.activation = activation self.approx_post = approx_post self.kl_method = kl_method self.n_mc_iter = n_mc_iter self.compute_kl = True super().__init__(in_channels, out_channels, kernel_size, stride, padding, dilation, True, groups, is_mixed_prior=is_mixed_prior, bias=bias) def forward(self, x): post = eval("vip." + self.approx_post) output, kl = post.forward(self, x, fun="conv_transpose1d") return output, kl class ConvTranspose2d(BayesLocScaleConvBaseBlock): """ Bayesian Convolution 2d layer. Learned parameters are mu and logsigmasq. """ def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, dilation=1, activation=None, is_mixed_prior=False, bias=False, approx_post="Radial", kl_method="repar", n_mc_iter=20, **kwargs): kernel_size = _pair(kernel_size) stride = _pair(stride) padding = _pair(padding) dilation = _pair(dilation) self.activation = activation self.approx_post = approx_post self.kl_method = kl_method self.n_mc_iter = n_mc_iter self.compute_kl = True super().__init__(in_channels, out_channels, kernel_size, stride, padding, dilation, True, groups, is_mixed_prior=is_mixed_prior, bias=bias) def forward(self, x): post = eval("vip." + self.approx_post) output, kl = post.forward(self, x, fun="conv_transpose2d") return output, kl class ConvTranspose3d(BayesLocScaleConvBaseBlock): """ Bayesian Convolution 3d layer. Learned parameters are mu and logsigmasq. """ def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, dilation=1, activation=None, is_mixed_prior=False, bias=False, approx_post="Radial", kl_method="repar", n_mc_iter=20, **kwargs): kernel_size = _triple(kernel_size) stride = _triple(stride) padding = _triple(padding) dilation = _triple(dilation) self.activation = activation self.approx_post = approx_post self.kl_method = kl_method self.n_mc_iter = n_mc_iter self.compute_kl = True super().__init__(in_channels, out_channels, kernel_size, stride, padding, dilation, True, groups, is_mixed_prior=is_mixed_prior, bias=bias) def forward(self, x): post = eval("vip." + self.approx_post) output, kl = post.forward(self, x, fun="conv_transpose3d") return output, kl class Linear(nn.Module): def __init__(self, in_features, out_features, prior_mu=0, prior_logsigmasq=0, bias=False, activation=None, approx_post="Radial", kl_method="repar", n_mc_iter=20, **kwargs): super(Linear, self).__init__() self.in_features = in_features self.out_features = out_features self.bias = bias self.activation = activation self.approx_post = approx_post self.kl_method = kl_method self.n_mc_iter = n_mc_iter self.compute_kl = True # Save parameters to fill prior self.prior_params = {} self.prior_params['mu'] = prior_mu self.prior_params['logsigmasq'] = prior_logsigmasq # Posterior parameters - trained: mu and logsigmasq self.post_mu = Parameter(torch.zeros(out_features, in_features)) self.post_logsigmasq = Parameter(torch.zeros(out_features, in_features)) if self.bias: self.post_mu_bias = Parameter(torch.zeros((out_features))) self.post_logsigmasq_bias = Parameter(torch.zeros((out_features))) self.reset_parameters() self.reset_priors() def reset_parameters(self): stdv = 1 / 2 # 1. / math.sqrt(self.post_mu.size(1)) self.post_mu.data.uniform_(-stdv, stdv) self.post_logsigmasq.data.uniform_(minvar, maxvar) # (0, 1) if self.bias: self.post_mu_bias.data.uniform_(-stdv, stdv) self.post_logsigmasq_bias.data.uniform_(minvar, maxvar) # (0, 1) def reset_priors(self): self.prior_mu = Parameter(torch.zeros_like(self.post_mu) + self.prior_params['mu'], requires_grad=False) self.prior_logsigmasq = Parameter( torch.zeros_like(self.post_logsigmasq) + self.prior_params['logsigmasq'], requires_grad=False) if self.bias: self.prior_mu_bias = Parameter( torch.zeros_like(self.post_mu_bias) + self.prior_params['mu'], requires_grad=False) self.prior_logsigmasq_bias = Parameter( torch.zeros_like(self.post_logsigmasq_bias) + self.prior_params['logsigmasq'], requires_grad=False) def extra_repr(self): # Displays arguments for a block s = ('{in_features}, {out_features}') s += ', bias = {bias}' if hasattr(self, "activation"): s += ', activation={activation}' if hasattr(self, "approx_post"): s += ', approx_post={approx_post}' if hasattr(self, "kl_method"): s += ', kl_method={kl_method}' if hasattr(self, "n_mc_iter"): s += ', n_mc_iter={n_mc_iter}' if hasattr(self, "compute_kl"): s += ', compute_kl={compute_kl}' return s.format(**self.__dict__) def forward(self, x): post = eval("vip." + self.approx_post) output, kl = post.forward(self, x, fun="linear") return output, kl
33.382456
81
0.504415
1,870
19,028
4.853476
0.094652
0.039665
0.020824
0.037021
0.825584
0.810269
0.771375
0.756611
0.74978
0.748127
0
0.008347
0.401881
19,028
569
82
33.441125
0.789122
0.070212
0
0.814732
0
0
0.0518
0.014376
0
0
0
0
0
1
0.046875
false
0
0.022321
0
0.107143
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
87488de1bc875a471cd59888fc9c4465319506ed
139
py
Python
cnct/client/models/__init__.py
ffaraone/connect-python-openapi-client
d38f51c96634b4fefa1d8d2edce8a2d96fa888f0
[ "Apache-2.0" ]
null
null
null
cnct/client/models/__init__.py
ffaraone/connect-python-openapi-client
d38f51c96634b4fefa1d8d2edce8a2d96fa888f0
[ "Apache-2.0" ]
null
null
null
cnct/client/models/__init__.py
ffaraone/connect-python-openapi-client
d38f51c96634b4fefa1d8d2edce8a2d96fa888f0
[ "Apache-2.0" ]
null
null
null
from cnct.client.models.base import Action, Collection, NS, Resource # noqa from cnct.client.models.resourceset import ResourceSet # noqa
69.5
76
0.805755
19
139
5.894737
0.631579
0.142857
0.25
0.357143
0
0
0
0
0
0
0
0
0.115108
139
2
77
69.5
0.910569
0.064748
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
5e498c18659f131e54ab98b2339fc9e124806690
135
py
Python
supabase_py/lib/__init__.py
bariqhibat/supabase-py
b20703d3d4117c911092212a796e53eb2f5286ca
[ "MIT" ]
null
null
null
supabase_py/lib/__init__.py
bariqhibat/supabase-py
b20703d3d4117c911092212a796e53eb2f5286ca
[ "MIT" ]
null
null
null
supabase_py/lib/__init__.py
bariqhibat/supabase-py
b20703d3d4117c911092212a796e53eb2f5286ca
[ "MIT" ]
null
null
null
from supabase_py.lib import auth_client, query_builder, realtime_client __all__ = ["auth_client", "query_builder", "realtime_client"]
33.75
71
0.807407
18
135
5.444444
0.611111
0.204082
0.306122
0.44898
0.734694
0.734694
0
0
0
0
0
0
0.088889
135
3
72
45
0.796748
0
0
0
0
0
0.288889
0
0
0
0
0
0
1
0
false
0
0.5
0
0.5
0
1
0
0
null
1
1
1
0
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
1
0
0
0
0
7
5e4c0a56e98d3890604e94406829c741ecf8e112
15,743
py
Python
src/tests/illumidesk/authenticators/test_setup_course_hook.py
jgwerner/illumidesk.next
a571b086e992e5f4562d3a605ca45b7ee4352def
[ "Apache-2.0" ]
null
null
null
src/tests/illumidesk/authenticators/test_setup_course_hook.py
jgwerner/illumidesk.next
a571b086e992e5f4562d3a605ca45b7ee4352def
[ "Apache-2.0" ]
2
2021-05-25T13:27:12.000Z
2021-05-25T13:33:34.000Z
src/tests/illumidesk/authenticators/test_setup_course_hook.py
jgwerner/illumidesk.next
a571b086e992e5f4562d3a605ca45b7ee4352def
[ "Apache-2.0" ]
2
2021-01-07T14:30:02.000Z
2021-04-28T11:49:42.000Z
import json import os from jupyterhub.auth import Authenticator import pytest from tornado.web import RequestHandler from tornado.httpclient import AsyncHTTPClient from unittest.mock import AsyncMock from unittest.mock import patch from illumidesk.apis.jupyterhub_api import JupyterHubAPI from illumidesk.apis.announcement_service import AnnouncementService from illumidesk.apis.nbgrader_service import NbGraderServiceHelper from illumidesk.authenticators.authenticator import LTI11Authenticator from illumidesk.authenticators.authenticator import LTI13Authenticator from illumidesk.authenticators.authenticator import setup_course_hook from illumidesk.authenticators.utils import LTIUtils @pytest.mark.asyncio async def test_setup_course_hook_is_assigned_to_lti11_authenticator_post_auth_hook(): """ Does the setup course hook get assigned to the post_auth_hook for the LTI11Authenticator? """ authenticator = LTI11Authenticator(post_auth_hook=setup_course_hook) assert authenticator.post_auth_hook == setup_course_hook @pytest.mark.asyncio async def test_setup_course_hook_is_assigned_to_lti13_authenticator_post_auth_hook(): """ Does the setup course hook get assigned to the post_auth_hook for the LTI13Authenticator? """ authenticator = LTI13Authenticator(post_auth_hook=setup_course_hook) assert authenticator.post_auth_hook == setup_course_hook @pytest.mark.asyncio() async def test_setup_course_hook_raises_environment_error_with_missing_org( monkeypatch, make_auth_state_dict, setup_course_hook_environ, make_mock_request_handler ): """ Is an environment error raised when the organization name is missing when calling the setup_course_hook function? """ monkeypatch.setenv('ORGANIZATION_NAME', '') local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict() with pytest.raises(EnvironmentError): await local_authenticator.post_auth_hook(local_authenticator, local_handler, local_authentication) @pytest.mark.asyncio() async def test_setup_course_hook_calls_normalize_strings( auth_state_dict, setup_course_environ, setup_course_hook_environ, make_mock_request_handler, make_http_response, mock_nbhelper, ): """ Does the setup_course_hook return normalized strings for the username and the course_id? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = auth_state_dict with patch.object(LTIUtils, 'normalize_string', return_value='intro101') as mock_normalize_string: with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None): with patch.object( AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request) ): _ = await setup_course_hook(local_authenticator, local_handler, local_authentication) assert mock_normalize_string.called @pytest.mark.asyncio() async def test_setup_course_hook_raises_json_decode_error_without_client_fetch_response( monkeypatch, setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_mock_request_handler, make_http_response, mock_nbhelper, ): """ Does the setup course hook raise a json decode error if the response form the setup course microservice is null or empty? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict() with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None): with patch.object( AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request, body=None) ): with pytest.raises(json.JSONDecodeError): await setup_course_hook(local_authenticator, local_handler, local_authentication) @pytest.mark.asyncio() async def test_setup_course_hook_calls_add_student_to_jupyterhub_group_when_role_is_learner( setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_http_response, make_mock_request_handler, mock_nbhelper, ): """ Is the jupyterhub_api add student to jupyterhub group function called when the user role is the learner role? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict() with patch.object( JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None ) as mock_add_student_to_jupyterhub_group: with patch.object(AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)): result = await setup_course_hook(local_authenticator, local_handler, local_authentication) assert mock_add_student_to_jupyterhub_group.called @patch('shutil.chown') @patch('pathlib.Path.mkdir') @patch('illumidesk.apis.nbgrader_service.Gradebook') @pytest.mark.asyncio() async def test_setup_course_hook_calls_add_user_to_nbgrader_gradebook_when_role_is_learner( mock_mkdir, mock_chown, mock_gradebook, monkeypatch, setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_mock_request_handler, make_http_response, ): """ Is the jupyterhub_api add user to nbgrader gradebook function called when the user role is the learner role? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict() with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None): with patch.object( NbGraderServiceHelper, 'add_user_to_nbgrader_gradebook', return_value=None ) as mock_add_user_to_nbgrader_gradebook: with patch.object( AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request) ): await setup_course_hook(local_authenticator, local_handler, local_authentication) assert mock_add_user_to_nbgrader_gradebook.called @pytest.mark.asyncio() async def test_setup_course_hook_calls_add_instructor_to_jupyterhub_group_when_role_is_instructor( monkeypatch, setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_mock_request_handler, make_http_response, mock_nbhelper, ): """ Is the jupyterhub_api add instructor to jupyterhub group function called when the user role is the instructor role? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict(user_role='Instructor') with patch.object( JupyterHubAPI, 'add_instructor_to_jupyterhub_group', return_value=None ) as mock_add_instructor_to_jupyterhub_group: with patch.object(AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)): await setup_course_hook(local_authenticator, local_handler, local_authentication) assert mock_add_instructor_to_jupyterhub_group.called @pytest.mark.asyncio() async def test_setup_course_hook_calls_add_instructor_to_jupyterhub_group_when_role_is_TeachingAssistant( monkeypatch, setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_mock_request_handler, make_http_response, mock_nbhelper, ): """ Is the jupyterhub_api add instructor to jupyterhub group function called when the user role is the instructor role? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict(user_role='urn:lti:role:ims/lis/TeachingAssistant') with patch.object( JupyterHubAPI, 'add_instructor_to_jupyterhub_group', return_value=None ) as mock_add_instructor_to_jupyterhub_group: with patch.object(AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)): await setup_course_hook(local_authenticator, local_handler, local_authentication) assert mock_add_instructor_to_jupyterhub_group.called @pytest.mark.asyncio() async def test_setup_course_hook_does_not_call_add_student_to_jupyterhub_group_when_role_is_instructor( setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_http_response, make_mock_request_handler, mock_nbhelper, ): """ Is the jupyterhub_api add student to jupyterhub group function called when the user role is the instructor role? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict(user_role='Instructor') with patch.object( JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None ) as mock_add_student_to_jupyterhub_group: with patch.object( JupyterHubAPI, 'add_instructor_to_jupyterhub_group', return_value=None ) as mock_add_instructor_to_jupyterhub_group: with patch.object( AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request) ): await setup_course_hook(local_authenticator, local_handler, local_authentication) assert not mock_add_student_to_jupyterhub_group.called assert mock_add_instructor_to_jupyterhub_group.called @pytest.mark.asyncio() async def test_setup_course_hook_does_not_call_add_instructor_to_jupyterhub_group_when_role_is_learner( setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_http_response, make_mock_request_handler, mock_nbhelper, ): """ Is the jupyterhub_api add instructor to jupyterhub group function not called when the user role is the learner role? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict() with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None): with patch.object( JupyterHubAPI, 'add_instructor_to_jupyterhub_group', return_value=None ) as mock_add_instructor_to_jupyterhub_group: with patch.object( AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request), ): await setup_course_hook(local_authenticator, local_handler, local_authentication) assert not mock_add_instructor_to_jupyterhub_group.called @pytest.mark.asyncio() async def test_setup_course_hook_initialize_data_dict( setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_http_response, make_mock_request_handler, mock_nbhelper, ): """ Is the data dictionary correctly initialized when properly setting the org env-var and and consistent with the course id value in the auth state? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict() expected_data = { 'org': 'test-org', 'course_id': 'intro101', 'domain': '127.0.0.1', } with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None): with patch.object(AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)): result = await setup_course_hook(local_authenticator, local_handler, local_authentication) assert expected_data['course_id'] == result['auth_state']['course_id'] assert expected_data['org'] == os.environ.get('ORGANIZATION_NAME') assert expected_data['domain'] == local_handler.request.host @pytest.mark.asyncio() async def test_setup_course_hook_calls_announcement_service_when_is_new_setup( setup_course_hook_environ, make_auth_state_dict, make_http_response, make_mock_request_handler, mock_nbhelper, ): """ Is the annuncement service called in new setup? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict() response_args = {'handler': local_handler.request, 'body': {'is_new_setup': True}} with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None): with patch.object( AsyncHTTPClient, 'fetch', side_effect=[ make_http_response(**response_args), None, ], # noqa: E231 ): AnnouncementService.add_announcement = AsyncMock(return_value=None) await setup_course_hook(local_authenticator, local_handler, local_authentication) assert AnnouncementService.add_announcement.called @pytest.mark.asyncio() async def test_is_new_course_initiates_rolling_update( setup_course_environ, setup_course_hook_environ, make_auth_state_dict, make_http_response, make_mock_request_handler, mock_nbhelper, ): """ If the course is a new setup does it initiate a rolling update? """ local_authenticator = Authenticator(post_auth_hook=setup_course_hook) local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator) local_authentication = make_auth_state_dict() response_args = {'handler': local_handler.request, 'body': {'is_new_setup': True}} with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None): with patch.object( AsyncHTTPClient, 'fetch', side_effect=[ make_http_response(**response_args), None, ], # noqa: E231 ) as mock_client: AnnouncementService.add_announcement = AsyncMock(return_value=None) await setup_course_hook(local_authenticator, local_handler, local_authentication) assert mock_client.called mock_client.assert_any_call( 'http://setup-course:8000/rolling-update', headers={'Content-Type': 'application/json'}, body='', method='POST', ) mock_client.assert_any_call( 'http://setup-course:8000', headers={'Content-Type': 'application/json'}, body='{"org": "test-org", "course_id": "intro101", "domain": "127.0.0.1"}', method='POST', )
40.574742
116
0.753478
1,878
15,743
5.887114
0.087859
0.07064
0.07869
0.047757
0.833936
0.81042
0.794863
0.784461
0.775868
0.763929
0
0.003949
0.179572
15,743
387
117
40.679587
0.852044
0.001334
0
0.74386
0
0.003509
0.071851
0.037422
0
0
0
0
0.059649
1
0
false
0
0.052632
0
0.052632
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
5ea732ea7b973e5911b7b86625fbf404f700de0e
35,207
py
Python
dataset.py
DeVriesMatt/pointMLP-pytorch
e9c09a2038551e83b072353f3fd7e3294463e892
[ "Apache-2.0" ]
null
null
null
dataset.py
DeVriesMatt/pointMLP-pytorch
e9c09a2038551e83b072353f3fd7e3294463e892
[ "Apache-2.0" ]
null
null
null
dataset.py
DeVriesMatt/pointMLP-pytorch
e9c09a2038551e83b072353f3fd7e3294463e892
[ "Apache-2.0" ]
null
null
null
import torch from torch.utils.data import Dataset from sklearn.preprocessing import LabelEncoder import pandas as pd import os from pyntcloud import PyntCloud import numpy as np import random def translate_pointcloud(pointcloud): xyz1 = np.random.uniform(low=2. / 3., high=3. / 2., size=[3]) xyz2 = np.random.uniform(low=-0.2, high=0.2, size=[3]) translated_pointcloud = np.add(np.multiply(pointcloud, xyz1), xyz2).astype('float32') return translated_pointcloud def jitter_pointcloud(pointcloud, sigma=1, clip=0.02): N, C = pointcloud.shape rotation = np.copy(pointcloud) rotation += np.clip(sigma * np.random.randn(N, C), -1 * clip, clip) return rotation def generate_24_rotations(): res = [] for id in [[0, 1, 2], [1, 2, 0], [2, 0, 1]]: R = np.identity(3)[:, id].astype(int) R1= np.asarray([R[:, 0], R[:, 1], R[:, 2]]).T R2 = np.asarray([-R[:, 0], -R[:, 1], R[:, 2]]).T R3 = np.asarray([-R[:, 0], R[:, 1], -R[:, 2]]).T R4 = np.asarray([R[:, 0], -R[:, 1], -R[:, 2]]).T res += [R1, R2, R3, R4] for id in [[0, 2, 1], [1, 0, 2], [2, 1, 0]]: R = np.identity(3)[:, id].astype(int) R1 = np.asarray([-R[:, 0], -R[:, 1], -R[:, 2]]).T R2 = np.asarray([-R[:, 0], R[:, 1], R[:, 2]]).T R3 = np.asarray([R[:, 0], -R[:, 1], R[:, 2]]).T R4 = np.asarray([R[:, 0], R[:, 1], -R[:, 2]]).T res += [R1, R2, R3, R4] return res def rotate_pointcloud(pointcloud): # theta = np.random.normal(0, (np.pi**2)/16, 1)[0] # print(theta) theta = np.pi * 2 * np.random.choice(24) / 24 rotation_matrix = np.array([[np.cos(theta), -np.sin(theta)], [np.sin(theta), np.cos(theta)]]) rotation = np.copy(pointcloud) rotation[:, [0, 2]] = pointcloud[:, [0, 2]].dot(rotation_matrix) # random rotation (x,z) return rotation, theta def three_d_rotation(pointcloud): alpha = np.pi * 2 * np.random.choice(24) / 24 beta = np.pi * 2 * np.random.choice(24) / 24 gamma = np.pi * 2 * np.random.choice(24) / 24 rotation_matrix = np.array( [[np.cos(beta) * np.cos(gamma), (np.sin(alpha) * np.sin(beta) * np.cos(gamma)) - (np.cos(alpha) * np.cos(gamma)), (np.cos(alpha) * np.sin(beta) * np.cos(gamma)) + (np.sin(alpha) * np.sin(gamma))], [np.cos(beta) * np.sin(gamma), (np.sin(alpha) * np.sin(beta) * np.sin(gamma)) + (np.cos(alpha) * np.cos(gamma)), (np.cos(alpha) * np.sin(beta) * np.sin(gamma)) - (np.sin(alpha) * np.cos(gamma))], [-np.sin(beta), np.sin(alpha) * np.cos(beta), np.cos(alpha) * np.cos(beta)]] ) rotation = np.copy(pointcloud) rotation[:, ] = pointcloud[:, ].dot(rotation_matrix) return rotation, (alpha, beta, gamma) class PointCloudDataset(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=False, ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.centring_only = centring_only self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] img_path = os.path.join( self.img_dir, treatment, self.new_df.loc[idx, "serialNumber"] ) image = PyntCloud.from_file(img_path + ".ply") image = torch.tensor(image.points.values) # TODO: take away after testing if self.centring_only: mean = torch.mean(image, 0) # mean = torch.tensor([[13.4828, 26.5144, 24.4187]]) # std = torch.tensor([[9.2821, 20.4512, 18.9049]]) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (image - mean) / std # / std # TODO: _____________________________________________ else: mean = torch.tensor([[13.4828, 26.5144, 24.4187]]) std = torch.tensor([[9.2821, 20.4512, 18.9049]]) image = (image - mean) / std # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) return image, label, feats class PointCloudDatasetAll(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=True, cell_component="cell", ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.centring_only = centring_only self.cell_component = cell_component self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) if self.cell_component == "cell": component_path = "stacked_pointcloud" else: component_path = "stacked_pointcloud_nucleus" img_path = os.path.join( self.img_dir, plate_num, component_path, treatment, self.new_df.loc[idx, "serialNumber"], ) image = PyntCloud.from_file(img_path + ".ply") image = torch.tensor(image.points.values) # TODO: take away after testing if self.centring_only: mean = torch.mean(image, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (image - mean) / std else: mean = torch.tensor([[13.4828, 26.5144, 24.4187]]) std = torch.tensor([[9.2821, 20.4512, 18.9049]]) image = (image - mean) / std # TODO: _____________________________________________ # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, label, serial_number class PointCloudDatasetAllBoth(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=False, cell_component="cell", proximal=1, ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.cell_component = cell_component self.proximal = proximal self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) & ( (self.annot_df.Treatment == "Nocodazole") | (self.annot_df.Treatment == "Blebbistatin") ) & (self.annot_df.Proximal == self.proximal) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) cell_path = "stacked_pointcloud" nuc_path = "stacked_pointcloud_nucleus" cell_img_path = os.path.join( self.img_dir, plate_num, cell_path, treatment, self.new_df.loc[idx, "serialNumber"], ) nuc_img_path = os.path.join( self.img_dir, plate_num, nuc_path, treatment, self.new_df.loc[idx, "serialNumber"], ) cell = PyntCloud.from_file(cell_img_path + ".ply") nuc = PyntCloud.from_file(nuc_img_path + ".ply") cell = torch.tensor(cell.points.values) nuc = torch.tensor(nuc.points.values) full = torch.tensor(np.concatenate((cell[:1024], nuc[:1024]))) mean = torch.mean(full, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (full - mean) / std # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) return image, label, feats class PointCloudDatasetAllBothNotSpec(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=False, cell_component="cell", proximal=1, ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.cell_component = cell_component self.proximal = proximal self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) cell_path = "stacked_pointcloud" nuc_path = "stacked_pointcloud_nucleus" cell_img_path = os.path.join( self.img_dir, plate_num, cell_path, treatment, self.new_df.loc[idx, "serialNumber"], ) nuc_img_path = os.path.join( self.img_dir, plate_num, nuc_path, treatment, self.new_df.loc[idx, "serialNumber"], ) cell = PyntCloud.from_file(cell_img_path + ".ply") nuc = PyntCloud.from_file(nuc_img_path + ".ply") cell = torch.tensor(cell.points.values) nuc = torch.tensor(nuc.points.values) full = torch.tensor(np.concatenate((cell[:1024], nuc[:1024]))) mean = torch.mean(full, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (full - mean) / std # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, label, serial_number class PointCloudDatasetAllBothNotSpec1024(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=True, cell_component="cell", proximal=1, ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.cell_component = cell_component self.proximal = proximal self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) cell_path = "stacked_pointcloud" nuc_path = "stacked_pointcloud_nucleus" cell_img_path = os.path.join( self.img_dir, plate_num, cell_path, treatment, self.new_df.loc[idx, "serialNumber"], ) nuc_img_path = os.path.join( self.img_dir, plate_num, nuc_path, treatment, self.new_df.loc[idx, "serialNumber"], ) cell = PyntCloud.from_file(cell_img_path + ".ply") nuc = PyntCloud.from_file(nuc_img_path + ".ply") cell = torch.tensor(cell.points.values) nuc = torch.tensor(nuc.points.values) full = torch.tensor(np.concatenate((cell[:512], nuc[:512]))) mean = torch.mean(full, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (full - mean) / std # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, label, serial_number class PointCloudDatasetAll1024(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=True, cell_component="cell", ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.centring_only = centring_only self.cell_component = cell_component self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) if self.cell_component == "cell": component_path = "stacked_pointcloud" else: component_path = "stacked_pointcloud_nucleus" img_path = os.path.join( self.img_dir, plate_num, component_path, treatment, self.new_df.loc[idx, "serialNumber"], ) image = PyntCloud.from_file(img_path + ".ply") image = torch.tensor(image.points.values) # TODO: take away after testing if self.centring_only: image = image[:1024] mean = torch.mean(image, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (image - mean) / std # TODO: _____________________________________________ # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, label, serial_number class PointCloudDatasetAllRotation1024(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=True, cell_component="cell", ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.centring_only = centring_only self.cell_component = cell_component self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) if self.cell_component == "cell": component_path = "stacked_pointcloud" else: component_path = "stacked_pointcloud_nucleus" img_path = os.path.join( self.img_dir, plate_num, component_path, treatment, self.new_df.loc[idx, "serialNumber"], ) image = PyntCloud.from_file(img_path + ".ply") image = torch.tensor(image.points.values) # TODO: take away after testing image = image[:1024] mean = torch.mean(image, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (image - mean) / std rotated_image, angles = three_d_rotation(image) # TODO: _____________________________________________ # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, rotated_image, angles, serial_number class PointCloudDatasetAllBothNotSpecRotation(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=False, cell_component="cell", proximal=1, ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.cell_component = cell_component self.proximal = proximal self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) cell_path = "stacked_pointcloud" nuc_path = "stacked_pointcloud_nucleus" cell_img_path = os.path.join( self.img_dir, plate_num, cell_path, treatment, self.new_df.loc[idx, "serialNumber"], ) nuc_img_path = os.path.join( self.img_dir, plate_num, nuc_path, treatment, self.new_df.loc[idx, "serialNumber"], ) cell = PyntCloud.from_file(cell_img_path + ".ply") nuc = PyntCloud.from_file(nuc_img_path + ".ply") cell = torch.tensor(cell.points.values) nuc = torch.tensor(nuc.points.values) full = torch.tensor(np.concatenate((cell[:1024], nuc[:1024]))) mean = torch.mean(full, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (full - mean) / std rotated_image, angles = three_d_rotation(image.numpy()) rotated_image = torch.tensor(rotated_image) angles = torch.tensor(angles) # TODO: _____________________________________________ # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, rotated_image, angles, serial_number class PointCloudDatasetAllBothNotSpecRotation1024(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=True, cell_component="cell", proximal=1, ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.cell_component = cell_component self.proximal = proximal self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) cell_path = "stacked_pointcloud" nuc_path = "stacked_pointcloud_nucleus" cell_img_path = os.path.join( self.img_dir, plate_num, cell_path, treatment, self.new_df.loc[idx, "serialNumber"], ) nuc_img_path = os.path.join( self.img_dir, plate_num, nuc_path, treatment, self.new_df.loc[idx, "serialNumber"], ) cell = PyntCloud.from_file(cell_img_path + ".ply") nuc = PyntCloud.from_file(nuc_img_path + ".ply") cell = torch.tensor(cell.points.values) nuc = torch.tensor(nuc.points.values) full = torch.tensor(np.concatenate((cell[:512], nuc[:512]))) mean = torch.mean(full, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (full - mean) / std rotated_image, angles = three_d_rotation(image.numpy()) rotated_image = torch.tensor(rotated_image) angles = torch.tensor(angles) # TODO: _____________________________________________ # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, rotated_image, angles, serial_number class PointCloudDatasetAllBothNotSpec2DRotation1024(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=True, cell_component="cell", proximal=1, ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.cell_component = cell_component self.proximal = proximal self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) cell_path = "stacked_pointcloud" nuc_path = "stacked_pointcloud_nucleus" cell_img_path = os.path.join( self.img_dir, plate_num, cell_path, treatment, self.new_df.loc[idx, "serialNumber"], ) nuc_img_path = os.path.join( self.img_dir, plate_num, nuc_path, treatment, self.new_df.loc[idx, "serialNumber"], ) cell = PyntCloud.from_file(cell_img_path + ".ply") nuc = PyntCloud.from_file(nuc_img_path + ".ply") cell = torch.tensor(cell.points.values) nuc = torch.tensor(nuc.points.values) full = torch.tensor(np.concatenate((cell[:512], nuc[:512]))) mean = torch.mean(full, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (full - mean) / std rotated_image, angles = rotate_pointcloud(image.numpy()) rotated_image = torch.tensor(rotated_image) angles = torch.tensor(angles) # TODO: _____________________________________________ # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, rotated_image, angles, serial_number class PointCloudDatasetAllBothKLDivergranceRotation1024(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=True, cell_component="cell", proximal=1, rotation_matrices=generate_24_rotations(), ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.cell_component = cell_component self.proximal = proximal self.rotation_matrices = rotation_matrices self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) cell_path = "stacked_pointcloud" nuc_path = "stacked_pointcloud_nucleus" cell_img_path = os.path.join( self.img_dir, plate_num, cell_path, treatment, self.new_df.loc[idx, "serialNumber"], ) nuc_img_path = os.path.join( self.img_dir, plate_num, nuc_path, treatment, self.new_df.loc[idx, "serialNumber"], ) cell = PyntCloud.from_file(cell_img_path + ".ply") nuc = PyntCloud.from_file(nuc_img_path + ".ply") cell = torch.tensor(cell.points.values) nuc = torch.tensor(nuc.points.values) full = torch.tensor(np.concatenate((cell[:512], nuc[:512]))) mean = torch.mean(full, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (full - mean) / std rotation_matrix = torch.tensor(self.rotation_matrices[random.randrange(0, 24)]).type(torch.FloatTensor) rotated_image = torch.matmul(image, rotation_matrix) # TODO: _____________________________________________ # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, rotated_image, serial_number class SimCLR1024Both(Dataset): def __init__( self, annotations_file, img_dir, img_size=400, label_col="Treatment", transform=None, target_transform=None, centring_only=True, cell_component="cell", proximal=1, rotation_matrices=generate_24_rotations(), ): self.annot_df = pd.read_csv(annotations_file) self.img_dir = img_dir self.img_size = img_size self.label_col = label_col self.transform = transform self.target_transform = target_transform self.cell_component = cell_component self.proximal = proximal self.rotation_matrices = rotation_matrices self.new_df = self.annot_df[ (self.annot_df.xDim <= self.img_size) & (self.annot_df.yDim <= self.img_size) & (self.annot_df.zDim <= self.img_size) ].reset_index(drop=True) # encode label le = LabelEncoder() label_col_enc = self.new_df.loc[:, self.label_col] label_col_enc = le.fit_transform(label_col_enc) self.new_df["label_col_enc"] = label_col_enc def __len__(self): return len(self.new_df) def __getitem__(self, idx): # read the image treatment = self.new_df.loc[idx, "Treatment"] plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"]) cell_path = "stacked_pointcloud" nuc_path = "stacked_pointcloud_nucleus" cell_img_path = os.path.join( self.img_dir, plate_num, cell_path, treatment, self.new_df.loc[idx, "serialNumber"], ) nuc_img_path = os.path.join( self.img_dir, plate_num, nuc_path, treatment, self.new_df.loc[idx, "serialNumber"], ) cell = PyntCloud.from_file(cell_img_path + ".ply") nuc = PyntCloud.from_file(nuc_img_path + ".ply") cell = torch.tensor(cell.points.values) nuc = torch.tensor(nuc.points.values) full = torch.tensor(np.concatenate((cell[:512], nuc[:512]))) mean = torch.mean(full, 0) std = torch.tensor([[20.0, 20.0, 20.0]]) image = (full - mean) / std rotation_matrix = torch.tensor( self.rotation_matrices[random.randrange(1, 24)]).type(torch.FloatTensor) rotated_image = torch.matmul(image, rotation_matrix) rotated_jitter = jitter_pointcloud(rotated_image) rotated_jitter_translated = translate_pointcloud(rotated_jitter) rotation_matrix2 = torch.tensor( self.rotation_matrices[random.randrange(1, 24)]).type(torch.FloatTensor) rotated_image2 = torch.matmul(rotated_image, rotation_matrix2) rotated_jitter2 = jitter_pointcloud(rotated_image2) rotated_jitter_translated2 = translate_pointcloud(rotated_jitter2) # TODO: _____________________________________________ # return encoded label as tensor label = self.new_df.loc[idx, "label_col_enc"] label = torch.tensor(label) # return the classical features as torch tensor feats = self.new_df.iloc[idx, 16:-4] feats = torch.tensor(feats) serial_number = self.new_df.loc[idx, "serialNumber"] return image, rotated_jitter_translated, rotated_jitter_translated2, serial_number
32.568918
111
0.594399
4,341
35,207
4.474315
0.04469
0.04505
0.057921
0.047572
0.918344
0.908871
0.908305
0.908305
0.901097
0.896514
0
0.021246
0.291448
35,207
1,080
112
32.599074
0.757356
0.057858
0
0.877698
0
0
0.051749
0.00864
0
0
0
0.000926
0
1
0.049161
false
0
0.009592
0.014388
0.107914
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
2196527ce4de91e80ddc8c31852b374dd4fbc1a6
63
py
Python
25/02/object.py
pylangstudy/201707
c1cc72667f1e0b6e8eef4ee85067d7fa4ca500b6
[ "CC0-1.0" ]
null
null
null
25/02/object.py
pylangstudy/201707
c1cc72667f1e0b6e8eef4ee85067d7fa4ca500b6
[ "CC0-1.0" ]
46
2017-06-30T22:19:07.000Z
2017-07-31T22:51:31.000Z
25/02/object.py
pylangstudy/201707
c1cc72667f1e0b6e8eef4ee85067d7fa4ca500b6
[ "CC0-1.0" ]
null
null
null
#class object print(object) print(object()) print(dir(object))
12.6
18
0.746032
9
63
5.222222
0.444444
0.702128
0.723404
0.93617
0
0
0
0
0
0
0
0
0.079365
63
4
19
15.75
0.810345
0.190476
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
0
0
0
1
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
0
0
0
1
0
7
21da568c1c2790d144346e4c561ee722bce0c779
9,389
py
Python
vip/stats/distances.py
VChristiaens/VIP2.7
92ce75f1004b4dd1480c3688124225ce8a98aca2
[ "MIT" ]
null
null
null
vip/stats/distances.py
VChristiaens/VIP2.7
92ce75f1004b4dd1480c3688124225ce8a98aca2
[ "MIT" ]
null
null
null
vip/stats/distances.py
VChristiaens/VIP2.7
92ce75f1004b4dd1480c3688124225ce8a98aca2
[ "MIT" ]
null
null
null
#! /usr/bin/env python """ Distance between images. """ from __future__ import division __author__ = 'C. Gomez @ ULg' __all__ = ['cube_distance', 'cube_distance_to_frame'] import numpy as np import scipy.stats from matplotlib import pyplot as plt from ..var import get_annulus def cube_distance(array, frame, mode='full', dist='sad', inradius=None, width=None, plot=True): """ Computes the distance (or similarity) between frames in a cube, using one as the reference. Depending on the mode, the whole image can be used, or just the pixels in a given annulus. The criteria used are: - the Manhattan distance (SAD or sum of absolute differences), - the Euclidean distance (square root of the sum of the squared differences), - the Mean Squared Error, - the Spearman correlation coefficient, - the Pearson correlation coefficient, - the Structural Similarity Index (SSIM). The SAD, MSE and Ecuclidean criteria are dissimilarity criteria, which means that 0 is perfect similarity. The Spearman and Pearson correlation coefficients, vary between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Parameters ---------- array : array_like Input cube or 3d array. frame : int Reference frame in the cube. mode : {'full','annulus'}, string optional Whether to use the full frames or a centered annulus. dist : {'sad','euclidean','mse','pearson','spearman'}, str optional Which criterion to use. inradius : None or int, optional The inner radius when mode is 'annulus'. width : None or int, optional The width when mode is 'annulus'. plot : {True, False}, bool optional Whether to plot the distances or not. Returns ------- lista : array_like 1d array of distances for each frame wrt the reference one. """ if not array.ndim ==3: raise TypeError('The input array is not a cube or 3d array') lista = [] n = array.shape[0] if mode=='full': frame_ref = array[frame] elif mode=='annulus' and inradius and width: frame_ref = get_annulus(array[frame], inradius, width, True) else: raise TypeError('Mode not recognized or missing parameters') for i in range(n): if mode=='full': framei = array[i] elif mode=='annulus': framei = get_annulus(array[i], inradius, width, True) if dist=='sad': lista.append(np.sum(abs(frame_ref - framei))) elif dist=='euclidean': lista.append(np.sqrt(np.sum((frame_ref - framei)**2))) elif dist=='mse': lista.append((np.sum((frame_ref - framei)**2))/len(frame_ref)) elif dist=='pearson': pears, _ = scipy.stats.pearsonr(frame_ref.ravel(), framei.ravel()) lista.append(pears) elif dist=='spearman': spear, _ = scipy.stats.spearmanr(frame_ref.ravel(), framei.ravel()) lista.append(spear) else: raise ValueError('Distance not recognized') lista = np.array(lista) median_cor = np.median(lista) mean_cor = np.mean(lista) if plot: _, ax = plt.subplots(figsize=(12,6)) x = range(len(lista)) ax.plot(x, lista, '-', color='blue', alpha=0.3) ax.plot(x, lista, '.', color='blue', alpha=0.5) ax.vlines(frame, ymin=np.nanmin(lista), ymax=np.nanmax(lista), colors='green', linestyles='dashed', lw=2, alpha=0.8, label='Frame '+str(frame)) ax.hlines(median_cor, xmin=-1, xmax=n+1, colors='purple', linestyles='solid', label='Median value : '+str(median_cor)) ax.hlines(mean_cor, xmin=-1, xmax=n+1, colors='red', linestyles='solid', label='Mean value : '+str(mean_cor)) plt.xlabel('Frame number') if dist=='sad': plt.ylabel('SAD - Manhattan distance') elif dist=='euclidean': plt.ylabel('Euclidean distance') elif dist=='pearson': plt.ylabel('Pearson correlation coefficient') elif dist=='spearman': plt.ylabel('Spearman correlation coefficient') elif dist=='mse': plt.ylabel('Mean squared error') elif dist=='ssim': plt.ylabel('Structural Similarity Index') plt.xlim(xmin=-1, xmax=n+1) plt.minorticks_on() plt.legend(fancybox=True, framealpha=0.5, fontsize=12, loc='best') plt.grid(which='both') return lista def cube_distance_to_frame(array, frame_ref, mode='full', dist='sad', inradius=None, width=None, plot=True): """ Computes the distance (or similarity) between frames in a cube and a reference image. Depending on the mode, the whole image can be used, or just the pixels in a given annulus. The criteria used are: - the Manhattan distance (SAD or sum of absolute differences), - the Euclidean distance (square root of the sum of the squared differences), - the Mean Squared Error, - the Spearman correlation coefficient, - the Pearson correlation coefficient, - the Structural Similarity Index (SSIM). The SAD, MSE and Ecuclidean criteria are dissimilarity criteria, which means that 0 is perfect similarity. The Spearman and Pearson correlation coefficients, vary between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. The Structural Similarity Index was proposed by Wang et al. 2004. (http://www.cns.nyu.edu/pub/eero/wang03-reprint.pdf) SSIM varies between -1 and 1, where 1 means perfect similarity. SSIM attempts to model the perceived change in the structural information of the image. Parameters ---------- array : array_like Input cube or 3d array. frame_ref : array_like Reference image. mode : {'full','annulus'}, string optional Whether to use the full frames or a centered annulus. dist : {'sad','euclidean','mse','pearson','spearman','ssim'}, str optional Which criterion to use. inradius : None or int, optional The inner radius when mode is 'annulus'. width : None or int, optional The width when mode is 'annulus'. plot : {True, False}, bool optional Whether to plot the distances or not. Returns ------- lista : array_like 1d array of distances for each frame wrt the reference one. """ if not array.ndim ==3: raise TypeError('The input array is not a cube or 3d array') lista = [] n = array.shape[0] if mode=='full': frame_ref = frame_ref elif mode=='annulus' and inradius and width: frame_ref = get_annulus(frame_ref, inradius, width, True) else: raise TypeError('Mode not recognized or missing parameters') for i in range(n): if mode=='full': framei = array[i] elif mode=='annulus': framei = get_annulus(array[i], inradius, width, True) if dist=='sad': lista.append(np.sum(abs(frame_ref - framei))) elif dist=='euclidean': lista.append(np.sqrt(np.sum((frame_ref - framei)**2))) elif dist=='mse': lista.append((np.sum((frame_ref - framei)**2))/len(frame_ref)) elif dist=='pearson': pears, _ = scipy.stats.pearsonr(frame_ref.ravel(), framei.ravel()) lista.append(pears) elif dist=='spearman': spear, _ = scipy.stats.spearmanr(frame_ref.ravel(), framei.ravel()) lista.append(spear) elif dist=='ssim': lista.append(ssim(frame_ref, framei, win_size=7, dynamic_range=frame_ref.max() - frame_ref.min())) else: raise ValueError('Distance not recognized') lista = np.array(lista) median_cor = np.median(lista) mean_cor = np.mean(lista) if plot: _, ax = plt.subplots(figsize=(12,6)) x = range(len(lista)) ax.plot(x, lista, '-', color='blue', alpha=0.3) ax.plot(x, lista, '.', color='blue', alpha=0.5) ax.hlines(median_cor, xmin=-1, xmax=n+1, colors='purple', linestyles='solid', label='Median value : '+str(median_cor)) ax.hlines(mean_cor, xmin=-1, xmax=n+1, colors='red', linestyles='solid', label='Mean value : '+str(mean_cor)) plt.xlabel('Frame number') if dist=='sad': plt.ylabel('SAD - Manhattan distance') elif dist=='euclidean': plt.ylabel('Euclidean distance') elif dist=='pearson': plt.ylabel('Pearson correlation coefficient') elif dist=='spearman': plt.ylabel('Spearman correlation coefficient') elif dist=='mse': plt.ylabel('Mean squared error') elif dist=='ssim': plt.ylabel('Structural Similarity Index') plt.xlim(xmin=-1, xmax=n+1) plt.minorticks_on() plt.legend(fancybox=True, framealpha=0.5, fontsize=12, loc='best') plt.grid(which='both') return lista
38.479508
81
0.604857
1,202
9,389
4.66223
0.190516
0.032834
0.017488
0.010707
0.8601
0.8601
0.8601
0.8601
0.8601
0.8601
0
0.010778
0.278624
9,389
243
82
38.63786
0.816625
0.335286
0
0.845588
0
0
0.155932
0.003729
0
0
0
0
0
1
0.014706
false
0
0.036765
0
0.066176
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
1d0dfb271ba5e998b37d3cae96c4833e8f0b11de
59
py
Python
testing.py
ChiragMakkar13/salary
9e761c497ba72622936db4618bb0b6f9911f0230
[ "MIT" ]
null
null
null
testing.py
ChiragMakkar13/salary
9e761c497ba72622936db4618bb0b6f9911f0230
[ "MIT" ]
null
null
null
testing.py
ChiragMakkar13/salary
9e761c497ba72622936db4618bb0b6f9911f0230
[ "MIT" ]
1
2022-02-21T15:31:18.000Z
2022-02-21T15:31:18.000Z
print("Welcome to bitbucket") print("Welcome to bitbucket")
29.5
29
0.779661
8
59
5.75
0.5
0.521739
0.608696
1
0
0
0
0
0
0
0
0
0.084746
59
2
30
29.5
0.851852
0
0
1
0
0
0.666667
0
0
0
0
0
0
1
0
true
0
0
0
0
1
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
1
0
0
1
0
null
0
0
0
0
0
0
1
0
0
0
0
1
0
9
df062286bbb9b5b0cb2a0e767d6553972b3e1f19
259
py
Python
mhealth/users/helpers.py
theCreedo/indigitous-mental-health-app
e3a66a6b2e7fbd105106d0315f78ea3c93647fe8
[ "MIT" ]
null
null
null
mhealth/users/helpers.py
theCreedo/indigitous-mental-health-app
e3a66a6b2e7fbd105106d0315f78ea3c93647fe8
[ "MIT" ]
null
null
null
mhealth/users/helpers.py
theCreedo/indigitous-mental-health-app
e3a66a6b2e7fbd105106d0315f78ea3c93647fe8
[ "MIT" ]
null
null
null
from werkzeug.security import generate_password_hash from werkzeug.security import check_password_hash def hash_pwd(password): return generate_password_hash(password) def check_password(hashed, password): return check_password_hash(hashed, password)
32.375
52
0.841699
34
259
6.117647
0.352941
0.230769
0.192308
0.25
0
0
0
0
0
0
0
0
0.104247
259
8
53
32.375
0.896552
0
0
0
1
0
0
0
0
0
0
0
0
1
0.333333
false
1
0.333333
0.333333
1
0
0
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
1
0
1
1
1
1
0
0
10
df151440c4b3219f5b8974dd48f9f511b4d6e7b7
192
py
Python
portfolio_project/views.py
madelinepet/portfolio
c9b1b6318a4ff0018cb55d722d4705057acfe1ee
[ "MIT" ]
null
null
null
portfolio_project/views.py
madelinepet/portfolio
c9b1b6318a4ff0018cb55d722d4705057acfe1ee
[ "MIT" ]
null
null
null
portfolio_project/views.py
madelinepet/portfolio
c9b1b6318a4ff0018cb55d722d4705057acfe1ee
[ "MIT" ]
null
null
null
from django.shortcuts import render def home_view(request): return render(request, 'generic/home.html') def projects_view(request): return render(request, 'generic/projects.html')
19.2
51
0.755208
25
192
5.72
0.52
0.153846
0.237762
0.321678
0.517483
0.517483
0
0
0
0
0
0
0.135417
192
9
52
21.333333
0.861446
0
0
0
0
0
0.197917
0.109375
0
0
0
0
0
1
0.4
false
0
0.2
0.4
1
0
1
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
0
1
1
0
0
7
df586da51d18420f97450825dd0bcc3906f70203
163
py
Python
unitorch/models/detectron2/backbone/__init__.py
fuliucansheng/UniTorch
47038321593ce4e7eabda555bd58c0cf89482146
[ "MIT" ]
2
2022-02-05T08:52:00.000Z
2022-03-27T07:01:34.000Z
unitorch/models/detectron2/backbone/__init__.py
Lixin-Qian/unitorch
47038321593ce4e7eabda555bd58c0cf89482146
[ "MIT" ]
null
null
null
unitorch/models/detectron2/backbone/__init__.py
Lixin-Qian/unitorch
47038321593ce4e7eabda555bd58c0cf89482146
[ "MIT" ]
1
2022-03-27T07:01:13.000Z
2022-03-27T07:01:13.000Z
# Copyright (c) FULIUCANSHENG. # Licensed under the MIT License. import unitorch.models.detectron2.backbone.swin import unitorch.models.detectron2.backbone.yolo5
27.166667
48
0.822086
20
163
6.7
0.75
0.208955
0.298507
0.447761
0.567164
0
0
0
0
0
0
0.02027
0.092025
163
5
49
32.6
0.885135
0.368098
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
8
10daadd1bc977eee2d42c2b629937df6494c1352
17,867
py
Python
snow/plugins/fix2_configuresnow.py
vPeteWalker/gts21
90d827028f5c8f21457de81c0b718f7adbe9a262
[ "MIT" ]
null
null
null
snow/plugins/fix2_configuresnow.py
vPeteWalker/gts21
90d827028f5c8f21457de81c0b718f7adbe9a262
[ "MIT" ]
null
null
null
snow/plugins/fix2_configuresnow.py
vPeteWalker/gts21
90d827028f5c8f21457de81c0b718f7adbe9a262
[ "MIT" ]
2
2021-04-06T16:59:32.000Z
2021-05-05T18:50:34.000Z
cat <<EOF > /tmp/ConfigureSNOW.py from selenium import webdriver from selenium.webdriver.firefox.options import Options from selenium.webdriver.common.keys import Keys from selenium.webdriver.common.action_chains import ActionChains import os, sys, time options = Options() options.headless = True profile = webdriver.FirefoxProfile() browser = webdriver.Firefox(options=options,firefox_profile=profile) # Service Now instance id need to be changes in the url snow_url = "https://@@{SNOWInstanceURL}@@/nav_to.do?" snow_username = "admin" snow_password = "@@{SNOWAdminPassword}@@" snow_users = "https://@@{SNOWInstanceURL}@@/sys_user_list.do?sysparm_userpref_module%3Dc5aa0fff0a0a0aa7009a39da035ea396%26sysparm_clear_stack%3Dtrue" snow_groups = "https://@@{SNOWInstanceURL}@@/sys_user_group_list.do?sysparm_userpref_module%3Dc5aa68730a0a0aa70036ced8b58ca05c%26sysparm_clear_stack%3Dtrue" user_criteria_plugin_url = "https://@@{SNOWInstanceURL}@@/\$allappsmgmt.do?sysparm_redirect%3Dtrue&sysparm_search=User%20Criteria%20Scoped%20API" glide_validate_property_url = "https://@@{SNOWInstanceURL}@@/sys_properties_list.do?sysparm_query=nameSTARTSWITHglide.sc.guide.tab.validate&sysparm_first_row=1&sysparm_view=&sysparm_choice_query_raw=&sysparm_list_header_search=true" glide_reset_cascade_url = "https://@@{SNOWInstanceURL}@@/sys_properties_list.do?sysparm_query=nameSTARTSWITHglide.sc.reset_cascade&sysparm_first_row=1&sysparm_view=&sysparm_choice_query_raw=&sysparm_list_header_search=true" tables_url = "https://@@{SNOWInstanceURL}@@/sys_db_object_list.do?sysparm_query=labelSTARTSWITHCatalog%20Client%20Scripts&sysparm_first_row=1&sysparm_view=&sysparm_choice_query_raw=&sysparm_list_header_search=true" sys_property = "sys_properties.list" property_name = "glide.sc.guide.tab.validate" ###Login to the ServiceNow Instance browser.get(snow_url) time.sleep(30) #frame = browser.find_element_by_xpath browser.switch_to.frame('gsft_main') time.sleep(5) browser.find_element_by_id("user_name").send_keys(snow_username) browser.find_element_by_id("user_password").send_keys(snow_password) browser.find_element_by_id("sysverb_login").click() print('Logged in to SNOW') browser.switch_to_default_content() #Create user browser.get(snow_users) time.sleep(30) browser.find_element_by_id("sysverb_new").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='sys_user.user_name']").send_keys("midserver") browser.find_element_by_xpath("//*[@id='sys_user.user_password']").send_keys(snow_password) browser.find_element_by_xpath("//*[@id='sysverb_insert']").click() time.sleep(5) browser.get(snow_users) time.sleep(10) browser.find_element_by_link_text("midserver").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='tabs2_list']/span[2]/span/span[2]").click() time.sleep(5) browser.find_element_by_id("sysverb_edit_m2m").click() time.sleep(5) browser.find_element_by_id("_sys_user_role").send_keys("mid_server") time.sleep(5) browser.find_element_by_xpath("//*[text()='mid_server']").click() time.sleep(5) browser.find_element_by_id("add_to_collection_button").click() time.sleep(5) browser.find_element_by_id("select_0_sysverb_save").click() time.sleep(5) browser.find_element_by_id("sysverb_update").click() time.sleep(15) print('midserver user account created') #User Criteria Scoped API Plugin Activation browser.get(user_criteria_plugin_url) time.sleep(45) browser.find_element_by_xpath("//*[@id='appSection']/div/render-plugin/div/div[1]/div[2]/div[2]/div/div/button").click() time.sleep(10) browser.switch_to.frame("pluginActivateFrame") time.sleep(5) browser.find_element_by_id("ok_button").click() browser.implicitly_wait(600) browser.find_element_by_id("sysparm_button_close").click() time.sleep(2) print('User Criteria Scoped API plugin activated') browser.switch_to_default_content() ###glide.sc.guide.tab.validate Property Change browser.get(glide_validate_property_url) time.sleep(30) browser.find_element_by_link_text("glide.sc.guide.tab.validate").click() time.sleep(10) browser.find_element_by_id("sys_properties.value").clear() browser.find_element_by_id("sys_properties.value").send_keys("true") browser.find_element_by_id("sysverb_update").click() time.sleep(15) print('glide.sc.guide.tab.validate updated') ###Table_Access_Script browser.get(snow_url) time.sleep(30) browser.find_element_by_id("filter").send_keys("sys.scripts.do") browser.find_element_by_id("filter").send_keys(Keys.ENTER) time.sleep(5) browser.switch_to.frame('gsft_main') time.sleep(5) text_area = browser.find_element_by_xpath("//*[@id='runscript']") text_area.send_keys( "var table_array_string='sys_user_has_role,sys_user_group,sys_user_grmember,sys_group_has_role,sc_category," "sc_catalog,item_option_new,catalog_ui_policy,catalog_script_client,user_criteria,question,question_choice,sysapproval_approver';") text_area.send_keys(Keys.ENTER) text_area.send_keys("var query_string='nameIN'+ table_array_string;") text_area.send_keys(Keys.ENTER) text_area.send_keys("var db_objectGR=new GlideRecord('sys_db_object');") text_area.send_keys(Keys.ENTER) text_area.send_keys("db_objectGR.addEncodedQuery(query_string);") text_area.send_keys(Keys.ENTER) text_area.send_keys("db_objectGR.query();") text_area.send_keys(Keys.ENTER) text_area.send_keys("while(db_objectGR.next())") text_area.send_keys(Keys.ENTER) text_area.send_keys("{") text_area.send_keys(Keys.ENTER) text_area.send_keys("gs.error(db_objectGR.label.getDisplayValue()+\"::\"+db_objectGR.read_access.getDisplayValue()+'-->');") text_area.send_keys(Keys.ENTER) text_area.send_keys("db_objectGR.read_access=true;") text_area.send_keys(Keys.ENTER) text_area.send_keys("db_objectGR.create_access=true;") text_area.send_keys(Keys.ENTER) text_area.send_keys("db_objectGR.update_access=true;") text_area.send_keys(Keys.ENTER) text_area.send_keys("db_objectGR.delete_access=false;") text_area.send_keys(Keys.ENTER) text_area.send_keys("db_objectGR.update();") text_area.send_keys(Keys.ENTER) text_area.send_keys("}") browser.find_element_by_xpath("/html/body/form/input[2]").click() time.sleep(15) print('Table permissions updated') browser.switch_to_default_content() ##Upload & Install build (v1.0+1.1) browser.get(snow_url) time.sleep(30) browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do") browser.find_element_by_id("filter").send_keys(Keys.ENTER) time.sleep(5) browser.switch_to.frame('gsft_main') time.sleep(5) browser.find_element_by_link_text("Import Update Set from XML").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Full Certified Build(v1.0+v1.1).xml") browser.implicitly_wait(10) browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click() browser.implicitly_wait(300) browser.find_element_by_link_text("Nutanix Calm").click() browser.implicitly_wait(5) browser.find_element_by_id("preview_update_set").click() browser.implicitly_wait(180) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() browser.implicitly_wait(10) try: browser.find_element_by_link_text("Skip remote update").click() browser.implicitly_wait(10) browser.find_element_by_link_text("Skip remote update").click() browser.implicitly_wait(10) browser.find_element_by_link_text("Skip remote update").click() browser.implicitly_wait(10) browser.find_element_by_link_text("Skip remote update").click() browser.implicitly_wait(10) browser.find_element_by_link_text("Skip remote update").click() browser.implicitly_wait(10) except: print('Skip remote update not there, committing update set') browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click() browser.implicitly_wait(600) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() browser.implicitly_wait(10) print('Calm v1.1 plugin installed') browser.switch_to_default_content() ######Upload & Install build (v1.2_5) browser.get(snow_url) time.sleep(30) browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do") browser.find_element_by_id("filter").send_keys(Keys.ENTER) time.sleep(5) browser.switch_to.frame('gsft_main') time.sleep(5) browser.find_element_by_link_text("Import Update Set from XML").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm V1.2_5.xml") browser.implicitly_wait(10) browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click() browser.implicitly_wait(300) browser.find_element_by_link_text("Nutanix Calm V1.2_5").click() browser.implicitly_wait(5) browser.find_element_by_id("preview_update_set").click() browser.implicitly_wait(90) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() browser.implicitly_wait(10) try: browser.find_element_by_link_text("Skip remote update").click() browser.implicitly_wait(10) except: print('Skip remote update not there') browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click() browser.implicitly_wait(600) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() print('Calm v1.2.5 plugin installed') browser.switch_to_default_content() ######Upload & Install build (v1.3_4) browser.get(snow_url) time.sleep(30) browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do") browser.find_element_by_id("filter").send_keys(Keys.ENTER) time.sleep(5) browser.switch_to.frame('gsft_main') time.sleep(5) browser.find_element_by_link_text("Import Update Set from XML").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm V1.3_4.xml") browser.implicitly_wait(10) browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click() browser.implicitly_wait(300) browser.find_element_by_link_text("Nutanix Calm V1.3_4").click() browser.implicitly_wait(5) browser.find_element_by_id("preview_update_set").click() browser.implicitly_wait(90) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() browser.implicitly_wait(10) browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click() browser.implicitly_wait(600) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() print('Calm v1.3.4 plugin installed') browser.switch_to_default_content() ###glide.sc.reset_cascade Property Change browser.get(glide_reset_cascade_url) time.sleep(30) browser.find_element_by_link_text("glide.sc.reset_cascade").click() time.sleep(10) browser.find_element_by_id("sys_properties.value").clear() browser.find_element_by_id("sys_properties.value").send_keys("true") browser.find_element_by_id("sysverb_update").click() time.sleep(15) print('glide.sc.reset_cascade updated') ###Table permission update #browser.get(tables_url) #time.sleep(30) #browser.find_element_by_link_text("Catalog Client Scripts").click() #time.sleep(5) #browser.find_element_by_xpath("//*[@id='tabs2_section']/span[3]/span[1]/span[2]").click() #time.sleep(5) #browser.find_element_by_xpath("//*[@id='label.ni.sys_db_object.update_access']").click() #time.sleep(5) #browser.find_element_by_id("sysverb_update").click() #time.sleep(5) #print('Catalog Client Scripts permissions updated') ######Upload & Install build (v1.4.22) browser.get(snow_url) time.sleep(30) browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do") browser.find_element_by_id("filter").send_keys(Keys.ENTER) time.sleep(5) browser.switch_to.frame('gsft_main') time.sleep(5) browser.find_element_by_link_text("Import Update Set from XML").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm V1.4.2.xml") browser.implicitly_wait(10) browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click() browser.implicitly_wait(300) browser.find_element_by_link_text("Nutanix Calm V1.4.2").click() browser.implicitly_wait(5) browser.find_element_by_id("preview_update_set").click() browser.implicitly_wait(90) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() browser.implicitly_wait(10) #browser.find_element_by_link_text("Skip remote update").click() #browser.implicitly_wait(10) browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click() browser.implicitly_wait(10) browser.find_element_by_xpath("//*[@id='ok_button']").click() browser.implicitly_wait(300) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() print('Calm v1.4.2 plugin installed') browser.switch_to_default_content() ######Upload & Install build (v1.4.3) browser.get(snow_url) time.sleep(30) browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do") browser.find_element_by_id("filter").send_keys(Keys.ENTER) time.sleep(5) browser.switch_to.frame('gsft_main') time.sleep(5) browser.find_element_by_link_text("Import Update Set from XML").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm V1.4.3.xml") browser.implicitly_wait(10) browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click() browser.implicitly_wait(300) browser.find_element_by_link_text("Nutanix Calm V1.4.3").click() browser.implicitly_wait(5) browser.find_element_by_id("preview_update_set").click() browser.implicitly_wait(90) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() browser.implicitly_wait(10) browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click() browser.implicitly_wait(300) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() print('Calm v1.4.3 plugin installed') browser.switch_to_default_content() ######Upload & Install "Nutanix Calm - User Approval" Workflow browser.get(snow_url) time.sleep(30) browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do") browser.find_element_by_id("filter").send_keys(Keys.ENTER) time.sleep(5) browser.switch_to.frame('gsft_main') time.sleep(5) browser.find_element_by_link_text("Import Update Set from XML").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm User Approval Workflow 1.0.xml") browser.implicitly_wait(10) browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click() browser.implicitly_wait(300) browser.find_element_by_link_text("User Approval Workflow").click() browser.implicitly_wait(5) browser.find_element_by_id("preview_update_set").click() browser.implicitly_wait(90) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() browser.implicitly_wait(10) browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click() browser.implicitly_wait(300) browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click() print('User Approval Workflow installed') browser.switch_to_default_content() #Create Calm Users group browser.get(snow_groups) time.sleep(30) browser.find_element_by_id("sysverb_new").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='sys_user_group.name']").send_keys("Calm Users") browser.find_element_by_xpath("//*[@id='sysverb_insert']").click() time.sleep(5) browser.find_element_by_link_text("Calm Users").click() time.sleep(5) browser.find_element_by_id("sysverb_edit_m2m").click() time.sleep(5) browser.find_element_by_id("_sys_user_role").send_keys("x_nuta2_nutanix_ca.user") time.sleep(5) browser.find_element_by_xpath("//*[text()='x_nuta2_nutanix_ca.user']").click() time.sleep(5) browser.find_element_by_id("add_to_collection_button").click() time.sleep(5) browser.find_element_by_id("select_0_sysverb_save").click() time.sleep(5) browser.find_element_by_id("sysverb_update").click() time.sleep(5) print('Calm Users group created') #Create Nutanix Calm users browser.get(snow_users) time.sleep(30) for x in range(1, 11): name = "operator{}".format('%02d' % x) browser.find_element_by_id("sysverb_new").click() time.sleep(5) browser.find_element_by_xpath("//*[@id='sys_user.user_name']").send_keys(name) browser.find_element_by_xpath("//*[@id='sys_user.user_password']").send_keys("nutanix/4u") browser.find_element_by_xpath("//*[@id='sysverb_insert']").click() time.sleep(5) browser.find_element_by_link_text(name).click() time.sleep(5) browser.find_element_by_xpath("//*[@id='tabs2_list']/span[3]/span/span[2]").click() time.sleep(3) browser.find_element_by_xpath("/html/body/div[2]/div[2]/div/div[3]/span/div[2]/div[1]/div/div[1]/button[3]").click() time.sleep(5) browser.find_element_by_id("_sys_user_group").send_keys("Calm Users") time.sleep(5) browser.find_element_by_xpath("//*[text()='Calm Users']").click() time.sleep(5) browser.find_element_by_id("add_to_collection_button").click() time.sleep(5) browser.find_element_by_id("select_0_sysverb_save").click() time.sleep(5) browser.find_element_by_id("sysverb_update").click() time.sleep(10) print('{} user account created'.format(name)) ######End browser.quit() EOF wget -P /tmp/ https://raw.githubusercontent.com/nutanix/Calm-Servicenow-Plugin/master/v1.1/Nutanix%20Full%20Certified%20Build\(v1.0%2Bv1.1\).xml wget -P /tmp/ https://raw.githubusercontent.com/nutanix/Calm-Servicenow-Plugin/master/v1.2/Nutanix%20Calm%20V1.2_5.xml wget -P /tmp/ https://raw.githubusercontent.com/nutanixworkshops/gts21/master/snow/plugins/Nutanix%20Calm%20V1.3_4.xml wget -P /tmp/ https://raw.githubusercontent.com/nutanixworkshops/gts21/master/snow/plugins/Nutanix%20Calm%20V1.4.2.xml wget -P /tmp/ https://raw.githubusercontent.com/nutanixworkshops/gts21/master/snow/plugins/Nutanix%20Calm%20V1.4.3.xml wget -P /tmp/ https://raw.githubusercontent.com/nutanixworkshops/gts21/master/snow/plugins/Nutanix%20Calm%20User%20Approval%20Workflow%201.0.xml python3 /tmp/ConfigureSNOW.py
44.334988
232
0.791347
2,761
17,867
4.805143
0.093444
0.103641
0.169594
0.188437
0.814804
0.779151
0.76174
0.742745
0.739956
0.716967
0
0.032987
0.046454
17,867
402
233
44.445274
0.74573
0.06095
0
0.675214
0
0.017094
0.330799
0.157318
0
0
0
0
0
0
null
null
0.011396
0.031339
null
null
0.045584
0
0
0
null
0
0
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
8
802ca0574914285367c8339e50cb2681ba6c5519
2,523
py
Python
posix_checkapi/TRACES/POT/ut_lind_net_socketoptions.py
JustinCappos/checkapi
2508c414869eda3479e1384b1bea65ec1e749d3b
[ "Apache-2.0" ]
null
null
null
posix_checkapi/TRACES/POT/ut_lind_net_socketoptions.py
JustinCappos/checkapi
2508c414869eda3479e1384b1bea65ec1e749d3b
[ "Apache-2.0" ]
null
null
null
posix_checkapi/TRACES/POT/ut_lind_net_socketoptions.py
JustinCappos/checkapi
2508c414869eda3479e1384b1bea65ec1e749d3b
[ "Apache-2.0" ]
null
null
null
import lind_test_server from lind_net_constants import * SyscallError = lind_test_server.SyscallError sockfd = lind_test_server.socket_syscall(AF_INET, SOCK_STREAM, 0) # should work... lind_test_server.bind_syscall(sockfd,'127.0.0.1',50102) # let's set some options!!! I'll check for function in different tests... assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==0) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==0) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==0) # reuseport assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==0) lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT, 1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==0) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==0) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1) # linger assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==0) lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER, 1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==1) # keepalive assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==0) lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE, 1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==1) # let's set some options!!! lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_SNDBUF, 1000) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_SNDBUF)==1000) lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_RCVBUF, 2000) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_RCVBUF)==2000) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==1) assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==1)
42.05
80
0.833928
378
2,523
5.171958
0.124339
0.126854
0.221995
0.303836
0.87468
0.856266
0.856266
0.846036
0.846036
0.809207
0
0.021501
0.059849
2,523
59
81
42.762712
0.802698
0.056282
0
0.625
0
0
0.003791
0
0
0
0
0
0.6875
1
0
false
0
0.0625
0
0.0625
0
0
0
0
null
0
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
1
0
0
0
0
0
0
0
0
0
9
802d2f8aac0966415a060edb9115f53510aa7679
22,771
py
Python
gasp/geometry.py
pgg1610/GASP-python
84fd6f2e57fb3b90dbccdec77dd2d2453c0fd884
[ "MIT" ]
42
2016-12-08T03:10:40.000Z
2022-03-16T09:50:33.000Z
gasp/geometry.py
pgg1610/GASP-python
84fd6f2e57fb3b90dbccdec77dd2d2453c0fd884
[ "MIT" ]
2
2017-10-06T02:44:12.000Z
2019-05-19T07:21:44.000Z
gasp/geometry.py
pgg1610/GASP-python
84fd6f2e57fb3b90dbccdec77dd2d2453c0fd884
[ "MIT" ]
28
2016-12-08T02:26:57.000Z
2022-02-14T06:41:57.000Z
# coding: utf-8 # Copyright (c) Henniggroup. # Distributed under the terms of the MIT License. from __future__ import division, unicode_literals, print_function """ Geometry module: This module contains classes to hold geometry-specific data and operations, including any additional constraints. All geometry classes must implement pad(), unpad() and get_size() methods. 1. Bulk: Data and operations for 3D bulk structures 2. Sheet: Data and operations for 2D sheet structures 3. Wire: Data and operations for 1D wire structures 4. Cluster: Data and operations for 0D cluster structures """ from pymatgen.core.lattice import Lattice from pymatgen.core.sites import Site import numpy as np class Bulk(object): ''' Contains data and operations specific to bulk structures (so not much...). ''' def __init__(self): ''' Makes a Bulk object. ''' self.shape = 'bulk' self.max_size = np.inf self.min_size = -np.inf self.padding = None def pad(self, cell, padding='from_geometry'): ''' Does nothing. Args: cell: the Cell to pad padding: the amount of vacuum padding to add. If set to 'from_geometry', then the value in self.padding is used. ''' pass def unpad(self, cell, constraints): ''' Does nothing. Args: cell: the Cell to unpad constraints: the Constraints of the search ''' pass def get_size(self, cell): ''' Returns 0. Args: cell: the Cell whose size to get ''' return 0 class Sheet(object): ''' Contains data and operations specific to sheet structures. ''' def __init__(self, geometry_parameters): ''' Makes a Sheet, and sets default parameter values if necessary. Args: geometry_parameters: a dictionary of parameters ''' self.shape = 'sheet' # default values self.default_max_size = np.inf self.default_min_size = -np.inf self.default_padding = 10 # parse the parameters, and set defaults if necessary # max size if 'max_size' not in geometry_parameters: self.max_size = self.default_max_size elif geometry_parameters['max_size'] in (None, 'default'): self.max_size = self.default_max_size else: self.max_size = geometry_parameters['max_size'] # min size if 'min_size' not in geometry_parameters: self.min_size = self.default_min_size elif geometry_parameters['min_size'] in (None, 'default'): self.min_size = self.default_min_size else: self.min_size = geometry_parameters['min_size'] # padding if 'padding' not in geometry_parameters: self.padding = self.default_padding elif geometry_parameters['padding'] in (None, 'default'): self.padding = self.default_padding else: self.padding = geometry_parameters['padding'] def pad(self, cell, padding='from_geometry'): ''' Modifies a cell by adding vertical vacuum padding and making the c-lattice vector normal to the plane of the sheet. The atoms are shifted to the center of the padded sheet. Args: cell: the Cell to pad padding: the amount of vacuum padding to add (in Angstroms). If not set, then the value in self.padding is used. ''' # get the padding amount if padding == 'from_geometry': pad_amount = self.padding else: pad_amount = padding # make the padded lattice cell.rotate_to_principal_directions() species = cell.species cartesian_coords = cell.cart_coords cart_bounds = cell.get_bounding_box(cart_coords=True) minz = cart_bounds[2][0] maxz = cart_bounds[2][1] layer_thickness = maxz - minz ax = cell.lattice.matrix[0][0] bx = cell.lattice.matrix[1][0] by = cell.lattice.matrix[1][1] padded_lattice = Lattice([[ax, 0.0, 0.0], [bx, by, 0.0], [0.0, 0.0, layer_thickness + pad_amount]]) # modify the cell to correspond to the padded lattice cell.modify_lattice(padded_lattice) site_indices = [] for i in range(len(cell.sites)): site_indices.append(i) cell.remove_sites(site_indices) for i in range(len(cartesian_coords)): cell.append(species[i], cartesian_coords[i], coords_are_cartesian=True) # translate the atoms back into the cell if needed, and shift them to # the vertical center cell.translate_atoms_into_cell() frac_bounds = cell.get_bounding_box(cart_coords=False) z_center = frac_bounds[2][0] + (frac_bounds[2][1] - frac_bounds[2][0])/2 translation_vector = [0, 0, 0.5 - z_center] site_indices = [i for i in range(len(cell.sites))] cell.translate_sites(site_indices, translation_vector, frac_coords=True, to_unit_cell=False) def unpad(self, cell, constraints): ''' Modifies a cell by removing vertical vacuum padding, leaving only enough to satisfy the per-species MID constraints, and makes the c-lattice vector normal to the plane of the sheet (if it isn't already). Args: cell: the Cell to unpad constraints: the Constraints of the search ''' # make the unpadded lattice cell.rotate_to_principal_directions() species = cell.species cartesian_coords = cell.cart_coords layer_thickness = self.get_size(cell) max_mid = constraints.get_max_mid() + 0.01 # just to be safe... ax = cell.lattice.matrix[0][0] bx = cell.lattice.matrix[1][0] by = cell.lattice.matrix[1][1] unpadded_lattice = Lattice([[ax, 0.0, 0.0], [bx, by, 0.0], [0.0, 0.0, layer_thickness + max_mid]]) # modify the cell to correspond to the unpadded lattice cell.modify_lattice(unpadded_lattice) site_indices = [] for i in range(len(cell.sites)): site_indices.append(i) cell.remove_sites(site_indices) for i in range(len(cartesian_coords)): cell.append(species[i], cartesian_coords[i], coords_are_cartesian=True) # translate the atoms back into the cell if needed, and shift them to # the vertical center cell.translate_atoms_into_cell() frac_bounds = cell.get_bounding_box(cart_coords=False) z_center = frac_bounds[2][0] + (frac_bounds[2][1] - frac_bounds[2][0])/2 translation_vector = [0, 0, 0.5 - z_center] site_indices = [i for i in range(len(cell.sites))] cell.translate_sites(site_indices, translation_vector, frac_coords=True, to_unit_cell=False) def get_size(self, cell): ''' Returns the layer thickness of a sheet structure, which is the maximum vertical distance between atoms in the cell. Precondition: the cell has already been put into sheet format (c lattice vector parallel to the z-axis and a and b lattice vectors in the x-y plane) Args: cell: the Cell whose size to get ''' cart_bounds = cell.get_bounding_box(cart_coords=True) layer_thickness = cart_bounds[2][1] - cart_bounds[2][0] return layer_thickness class Wire(object): ''' Contains data and operations specific to wire structures. ''' def __init__(self, geometry_parameters): ''' Makes a Wire, and sets default parameter values if necessary. Args: geometry_parameters: a dictionary of parameters ''' self.shape = 'wire' # default values self.default_max_size = np.inf self.default_min_size = -np.inf self.default_padding = 10 # parse the parameters, and set defaults if necessary # max size if 'max_size' not in geometry_parameters: self.max_size = self.default_max_size elif geometry_parameters['max_size'] in (None, 'default'): self.max_size = self.default_max_size else: self.max_size = geometry_parameters['max_size'] # min size if 'min_size' not in geometry_parameters: self.min_size = self.default_min_size elif geometry_parameters['min_size'] in (None, 'default'): self.min_size = self.default_min_size else: self.min_size = geometry_parameters['min_size'] # padding if 'padding' not in geometry_parameters: self.padding = self.default_padding elif geometry_parameters['padding'] in (None, 'default'): self.padding = self.default_padding else: self.padding = geometry_parameters['padding'] def pad(self, cell, padding='from_geometry'): ''' Modifies a cell by making the c lattice vector parallel to z-axis, and adds vacuum padding around the structure in the x and y directions by replacing a and b lattice vectors with padded vectors along the x and y axes, respectively. The atoms are shifted to the center of the padded cell. Args: cell: the Cell to pad padding: the amount of vacuum padding to add (in Angstroms). If not set, then the value in self.padding is used. ''' # get the padding amount if padding == 'from_geometry': pad_amount = self.padding else: pad_amount = padding # make the padded lattice cell.rotate_c_parallel_to_z() species = cell.species cartesian_coords = cell.cart_coords cart_bounds = cell.get_bounding_box(cart_coords=True) x_min = cart_bounds[0][0] x_max = cart_bounds[0][1] y_min = cart_bounds[1][0] y_max = cart_bounds[1][1] x_extent = x_max - x_min y_extent = y_max - y_min cz = cell.lattice.matrix[2][2] padded_lattice = Lattice([[x_extent + pad_amount, 0, 0], [0, y_extent + pad_amount, 0], [0, 0, cz]]) # modify the cell to correspond to the padded lattice cell.modify_lattice(padded_lattice) site_indices = [] for i in range(len(cell.sites)): site_indices.append(i) cell.remove_sites(site_indices) for i in range(len(cartesian_coords)): cell.append(species[i], cartesian_coords[i], coords_are_cartesian=True) # translate the atoms back into the cell if needed, and shift them to # the horizontal center cell.translate_atoms_into_cell() frac_bounds = cell.get_bounding_box(cart_coords=False) x_center = frac_bounds[0][0] + (frac_bounds[0][1] - frac_bounds[0][0])/2 y_center = frac_bounds[1][0] + (frac_bounds[1][1] - frac_bounds[1][0])/2 translation_vector = [0.5 - x_center, 0.5 - y_center, 0.0] site_indices = [i for i in range(len(cell.sites))] cell.translate_sites(site_indices, translation_vector, frac_coords=True, to_unit_cell=False) def unpad(self, cell, constraints): ''' Modifies a cell by removing horizontal vacuum padding around a wire, leaving only enough to satisfy the per-species MID constraints, and makes the three lattice vectors lie along the three Cartesian directions. Args: cell: the Cell to unpad constraints: the Constraints of the search ''' # make the unpadded lattice cell.rotate_c_parallel_to_z() species = cell.species cartesian_coords = cell.cart_coords cart_bounds = cell.get_bounding_box(cart_coords=True) x_min = cart_bounds[0][0] x_max = cart_bounds[0][1] y_min = cart_bounds[1][0] y_max = cart_bounds[1][1] x_extent = x_max - x_min y_extent = y_max - y_min cz = cell.lattice.matrix[2][2] max_mid = constraints.get_max_mid() + 0.01 # just to be safe... unpadded_lattice = Lattice([[x_extent + max_mid, 0.0, 0.0], [0, y_extent + max_mid, 0.0], [0.0, 0.0, cz]]) # modify the cell to correspond to the unpadded lattice cell.modify_lattice(unpadded_lattice) site_indices = [] for i in range(len(cell.sites)): site_indices.append(i) cell.remove_sites(site_indices) for i in range(len(cartesian_coords)): cell.append(species[i], cartesian_coords[i], coords_are_cartesian=True) # translate the atoms back into the cell if needed, and shift them to # the horizontal center cell.translate_atoms_into_cell() frac_bounds = cell.get_bounding_box(cart_coords=False) x_center = frac_bounds[0][0] + (frac_bounds[0][1] - frac_bounds[0][0])/2 y_center = frac_bounds[1][0] + (frac_bounds[1][1] - frac_bounds[1][0])/2 translation_vector = [0.5 - x_center, 0.5 - y_center, 0.0] site_indices = [i for i in range(len(cell.sites))] cell.translate_sites(site_indices, translation_vector, frac_coords=True, to_unit_cell=False) def get_size(self, cell): ''' Returns the diameter of a wire structure, defined as the maximum distance between atoms projected to the x-y plane. Precondition: the cell has already been put into wire format (c lattice vector is parallel to z-axis and a and b lattice vectors in the x-y plane), and all sites are located inside the cell (i.e., have fractional coordinates between 0 and 1). Args: cell: the Cell whose size to get ''' max_distance = 0 for site_i in cell.sites: # make Site versions of each PeriodicSite so that the computed # distance won't include periodic images non_periodic_site_i = Site(site_i.species_and_occu, [site_i.coords[0], site_i.coords[1], 0.0]) for site_j in cell.sites: non_periodic_site_j = Site(site_j.species_and_occu, [site_j.coords[0], site_j.coords[1], 0.0]) distance = non_periodic_site_i.distance(non_periodic_site_j) if distance > max_distance: max_distance = distance return max_distance class Cluster(object): ''' Contains data and operations specific to clusters. ''' def __init__(self, geometry_parameters): ''' Makes a Cluster, and sets default parameter values if necessary. Args: geometry_parameters: a dictionary of parameters ''' self.shape = 'cluster' # default values self.default_max_size = np.inf self.default_min_size = -np.inf self.default_padding = 10 # parse the parameters, and set defaults if necessary # max size if 'max_size' not in geometry_parameters: self.max_size = self.default_max_size elif geometry_parameters['max_size'] in (None, 'default'): self.max_size = self.default_max_size else: self.max_size = geometry_parameters['max_size'] # min size if 'min_size' not in geometry_parameters: self.min_size = self.default_min_size elif geometry_parameters['min_size'] in (None, 'default'): self.min_size = self.default_min_size else: self.min_size = geometry_parameters['min_size'] # padding if 'padding' not in geometry_parameters: self.padding = self.default_padding elif geometry_parameters['padding'] in (None, 'default'): self.padding = self.default_padding else: self.padding = geometry_parameters['padding'] def pad(self, cell, padding='from_geometry'): ''' Modifies a cell by replacing the three lattice vectors with ones along the three Cartesian directions and adding vacuum padding to each one. The atoms are shifted to the center of the padded cell. Args: cell: the Cell to pad padding: the amount of vacuum padding to add (in Angstroms). If not set, then the value in self.padding is used. ''' # get the padding amount if padding == 'from_geometry': pad_amount = self.padding else: pad_amount = padding # make the padded lattice species = cell.species cartesian_coords = cell.cart_coords cart_bounds = cell.get_bounding_box(cart_coords=True) x_min = cart_bounds[0][0] x_max = cart_bounds[0][1] y_min = cart_bounds[1][0] y_max = cart_bounds[1][1] z_min = cart_bounds[2][0] z_max = cart_bounds[2][1] x_extent = x_max - x_min y_extent = y_max - y_min z_extent = z_max - z_min padded_lattice = Lattice([[x_extent + pad_amount, 0, 0], [0, y_extent + pad_amount, 0], [0, 0, z_extent + pad_amount]]) # modify the cell to correspond to the padded lattice cell.modify_lattice(padded_lattice) site_indices = [] for i in range(len(cell.sites)): site_indices.append(i) cell.remove_sites(site_indices) for i in range(len(cartesian_coords)): cell.append(species[i], cartesian_coords[i], coords_are_cartesian=True) # translate the atoms back into the cell if needed, and shift them to # the center cell.translate_atoms_into_cell() frac_bounds = cell.get_bounding_box(cart_coords=False) x_center = frac_bounds[0][0] + (frac_bounds[0][1] - frac_bounds[0][0])/2 y_center = frac_bounds[1][0] + (frac_bounds[1][1] - frac_bounds[1][0])/2 z_center = frac_bounds[2][0] + (frac_bounds[2][1] - frac_bounds[2][0])/2 translation_vector = [0.5 - x_center, 0.5 - y_center, 0.5 - z_center] site_indices = [i for i in range(len(cell.sites))] cell.translate_sites(site_indices, translation_vector, frac_coords=True, to_unit_cell=False) def unpad(self, cell, constraints): ''' Modifies a cell by removing vacuum padding in every direction, leaving only enough to satisfy the per-species MID constraints, and makes the three lattice vectors lie along the three Cartesian directions. Args: cell: the Cell to unpad constraints: the Constraints of the search ''' # make the unpadded lattice species = cell.species cartesian_coords = cell.cart_coords cart_bounds = cell.get_bounding_box(cart_coords=True) x_min = cart_bounds[0][0] x_max = cart_bounds[0][1] y_min = cart_bounds[1][0] y_max = cart_bounds[1][1] z_min = cart_bounds[2][0] z_max = cart_bounds[2][1] x_extent = x_max - x_min y_extent = y_max - y_min z_extent = z_max - z_min max_mid = constraints.get_max_mid() + 0.01 # just to be safe... unpadded_lattice = Lattice([[x_extent + max_mid, 0.0, 0.0], [0, y_extent + max_mid, 0.0], [0.0, 0.0, z_extent + max_mid]]) # modify the cell to correspond to the unpadded lattice cell.modify_lattice(unpadded_lattice) site_indices = [] for i in range(len(cell.sites)): site_indices.append(i) cell.remove_sites(site_indices) for i in range(len(cartesian_coords)): cell.append(species[i], cartesian_coords[i], coords_are_cartesian=True) # translate the atoms back into the cell if needed, and shift them to # the center cell.translate_atoms_into_cell() frac_bounds = cell.get_bounding_box(cart_coords=False) x_center = frac_bounds[0][0] + (frac_bounds[0][1] - frac_bounds[0][0])/2 y_center = frac_bounds[1][0] + (frac_bounds[1][1] - frac_bounds[1][0])/2 z_center = frac_bounds[2][0] + (frac_bounds[2][1] - frac_bounds[2][0])/2 translation_vector = [0.5 - x_center, 0.5 - y_center, 0.5 - z_center] site_indices = [i for i in range(len(cell.sites))] cell.translate_sites(site_indices, translation_vector, frac_coords=True, to_unit_cell=False) def get_size(self, cell): ''' Returns the diameter of a cluster structure, defined as the maximum distance between atoms in the cell. Precondition: all sites are located inside the cell (i.e., have fractional coordinates between 0 and 1) Args: cell: the Cell whose size to get ''' max_distance = 0 for site_i in cell.sites: # make Site versions of each PeriodicSite so that the computed # distance won't include periodic images non_periodic_site_i = Site(site_i.species_and_occu, site_i.coords) for site_j in cell.sites: non_periodic_site_j = Site(site_j.species_and_occu, site_j.coords) distance = non_periodic_site_i.distance(non_periodic_site_j) if distance > max_distance: max_distance = distance return max_distance
36.846278
79
0.58715
2,942
22,771
4.344324
0.080557
0.010015
0.007511
0.005633
0.892262
0.876223
0.874266
0.850481
0.822706
0.819419
0
0.017957
0.329893
22,771
617
80
36.905997
0.819647
0.247156
0
0.885802
0
0
0.024317
0
0
0
0
0
0
1
0.049383
false
0.006173
0.012346
0
0.08642
0.003086
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
8042bea693f001a0eec6ce6dadeba016db356440
5,770
py
Python
a10sdk/core/cgnv6/cgnv6_stateful_firewall_alg_sip.py
deepfield/a10sdk-python
bfaa58099f51f085d5e91652d1d1a3fd5c529d5d
[ "Apache-2.0" ]
16
2015-05-20T07:26:30.000Z
2021-01-23T11:56:57.000Z
a10sdk/core/cgnv6/cgnv6_stateful_firewall_alg_sip.py
deepfield/a10sdk-python
bfaa58099f51f085d5e91652d1d1a3fd5c529d5d
[ "Apache-2.0" ]
6
2015-03-24T22:07:11.000Z
2017-03-28T21:31:18.000Z
a10sdk/core/cgnv6/cgnv6_stateful_firewall_alg_sip.py
deepfield/a10sdk-python
bfaa58099f51f085d5e91652d1d1a3fd5c529d5d
[ "Apache-2.0" ]
23
2015-03-29T15:43:01.000Z
2021-06-02T17:12:01.000Z
from a10sdk.common.A10BaseClass import A10BaseClass class SamplingEnable(A10BaseClass): """This class does not support CRUD Operations please use parent. :param counters1: {"enum": ["all", "stat-request", "stat-response", "method-register", "method-invite", "method-ack", "method-cancel", "method-bye", "method-port-config", "method-prack", "method-subscribe", "method-notify", "method-publish", "method-info", "method-refer", "method-message", "method-update", "method-unknown", "parse-error", "keep-alive", "contact-error", "sdp-error", "rtp-port-no-op", "rtp-rtcp-port-success", "rtp-port-failure", "rtcp-port-failure", "contact-port-no-op", "contact-port-success", "contact-port-failure", "contact-new", "contact-alloc-failure", "contact-eim", "contact-eim-set", "rtp-new", "rtp-alloc-failure", "rtp-eim"], "type": "string", "description": "'all': all; 'stat-request': Request Received; 'stat-response': Response Received; 'method-register': Method REGISTER; 'method-invite': Method INVITE; 'method-ack': Method ACK; 'method-cancel': Method CANCEL; 'method-bye': Method BYE; 'method-port-config': Method OPTIONS; 'method-prack': Method PRACK; 'method-subscribe': Method SUBSCRIBE; 'method-notify': Method NOTIFY; 'method-publish': Method PUBLISH; 'method-info': Method INFO; 'method-refer': Method REFER; 'method-message': Method MESSAGE; 'method-update': Method UPDATE; 'method-unknown': Method Unknown; 'parse-error': Message Parse Error; 'keep-alive': Keep Alive; 'contact-error': Contact Process Error; 'sdp-error': SDP Process Error; 'rtp-port-no-op': RTP Port No Op; 'rtp-rtcp-port-success': RTP RTCP Port Success; 'rtp-port-failure': RTP Port Failure; 'rtcp-port-failure': RTCP Port Failure; 'contact-port-no-op': Contact Port No Op; 'contact-port-success': Contact Port Success; 'contact-port-failure': Contact Port Failure; 'contact-new': Contact Alloc; 'contact-alloc-failure': Contact Alloc Failure; 'contact-eim': Contact EIM; 'contact-eim-set': Contact EIM Set; 'rtp-new': RTP Alloc; 'rtp-alloc-failure': RTP Alloc Failure; 'rtp-eim': RTP EIM; ", "format": "enum"} :param DeviceProxy: The device proxy for REST operations and session handling. Refer to `common/device_proxy.py` """ def __init__(self, **kwargs): self.ERROR_MSG = "" self.b_key = "sampling-enable" self.DeviceProxy = "" self.counters1 = "" for keys, value in kwargs.items(): setattr(self,keys, value) class Sip(A10BaseClass): """ :param sampling_enable: {"minItems": 1, "items": {"type": "object"}, "uniqueItems": true, "type": "array", "array": [{"properties": {"optional": true, "counters1": {"enum": ["all", "stat-request", "stat-response", "method-register", "method-invite", "method-ack", "method-cancel", "method-bye", "method-port-config", "method-prack", "method-subscribe", "method-notify", "method-publish", "method-info", "method-refer", "method-message", "method-update", "method-unknown", "parse-error", "keep-alive", "contact-error", "sdp-error", "rtp-port-no-op", "rtp-rtcp-port-success", "rtp-port-failure", "rtcp-port-failure", "contact-port-no-op", "contact-port-success", "contact-port-failure", "contact-new", "contact-alloc-failure", "contact-eim", "contact-eim-set", "rtp-new", "rtp-alloc-failure", "rtp-eim"], "type": "string", "description": "'all': all; 'stat-request': Request Received; 'stat-response': Response Received; 'method-register': Method REGISTER; 'method-invite': Method INVITE; 'method-ack': Method ACK; 'method-cancel': Method CANCEL; 'method-bye': Method BYE; 'method-port-config': Method OPTIONS; 'method-prack': Method PRACK; 'method-subscribe': Method SUBSCRIBE; 'method-notify': Method NOTIFY; 'method-publish': Method PUBLISH; 'method-info': Method INFO; 'method-refer': Method REFER; 'method-message': Method MESSAGE; 'method-update': Method UPDATE; 'method-unknown': Method Unknown; 'parse-error': Message Parse Error; 'keep-alive': Keep Alive; 'contact-error': Contact Process Error; 'sdp-error': SDP Process Error; 'rtp-port-no-op': RTP Port No Op; 'rtp-rtcp-port-success': RTP RTCP Port Success; 'rtp-port-failure': RTP Port Failure; 'rtcp-port-failure': RTCP Port Failure; 'contact-port-no-op': Contact Port No Op; 'contact-port-success': Contact Port Success; 'contact-port-failure': Contact Port Failure; 'contact-new': Contact Alloc; 'contact-alloc-failure': Contact Alloc Failure; 'contact-eim': Contact EIM; 'contact-eim-set': Contact EIM Set; 'rtp-new': RTP Alloc; 'rtp-alloc-failure': RTP Alloc Failure; 'rtp-eim': RTP EIM; ", "format": "enum"}}}]} :param sip_value: {"optional": true, "enum": ["disable"], "type": "string", "description": "'disable': Disable ALG; ", "format": "enum"} :param uuid: {"description": "uuid of the object", "format": "string", "minLength": 1, "modify-not-allowed": 1, "optional": true, "maxLength": 64, "type": "string"} :param DeviceProxy: The device proxy for REST operations and session handling. Refer to `common/device_proxy.py` Class Description:: Configure SIP ALG for NAT stateful firewall (default: enabled). Class sip supports CRUD Operations and inherits from `common/A10BaseClass`. This class is the `"PARENT"` class for this module.` URL for this object:: `https://<Hostname|Ip address>//axapi/v3/cgnv6/stateful-firewall/alg/sip`. """ def __init__(self, **kwargs): self.ERROR_MSG = "" self.required=[] self.b_key = "sip" self.a10_url="/axapi/v3/cgnv6/stateful-firewall/alg/sip" self.DeviceProxy = "" self.sampling_enable = [] self.sip_value = "" self.uuid = "" for keys, value in kwargs.items(): setattr(self,keys, value)
93.064516
2,088
0.685615
751
5,770
5.242344
0.167776
0.050292
0.024384
0.016764
0.804928
0.804928
0.804928
0.787656
0.770892
0.770892
0
0.005215
0.135875
5,770
61
2,089
94.590164
0.784396
0.857539
0
0.47619
0
0
0.077326
0.053735
0
0
0
0
0
1
0.095238
false
0
0.047619
0
0.238095
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
1
0
0
1
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
339175e00f73565858eacb3fb6d3de894fac5da4
17,785
py
Python
tests/test_libcloud_extended.py
antoinebourayne/sd2c
c76a0c56d5836caba9e6b90cdf7235516e2dd694
[ "Apache-2.0" ]
null
null
null
tests/test_libcloud_extended.py
antoinebourayne/sd2c
c76a0c56d5836caba9e6b90cdf7235516e2dd694
[ "Apache-2.0" ]
null
null
null
tests/test_libcloud_extended.py
antoinebourayne/sd2c
c76a0c56d5836caba9e6b90cdf7235516e2dd694
[ "Apache-2.0" ]
null
null
null
import logging import os from unittest import TestCase import coloredlogs from dotenv import dotenv_values, find_dotenv import unittest.mock import sd2c as sshcrosscloud from sd2c import utils from sd2c.libcloud_extended import ProviderSpecific, SpecificAWS, get_provider_specific_driver from sd2c.ssh_cross_cloud import SSHCrossCloud import libcloud.compute.drivers.ec2 import libcloud.compute.drivers.azure_arm import libcloud.compute.base import libcloud.compute.providers def side_effect_create_local_rsa_key_pair_no_key(path): if path == "a": return False if path == "a.pub": return False def side_effect_create_local_rsa_key_pair_private(path): if path == "a": return True if path == "a.pub": return False class FakeProviderDriver: def __init__(self): self.node_id = None self.volume_name = None self.list_nodes_empty = False self.node_state = None self.terminate = True self.stop = True self.start = True self.list_nodes_called = False self.ex_stop_node_called = False self.ex_start_node_called = False self.destroy_node_called = False self.destroy_volume_called = False def list_nodes(self, arg=None): self.list_nodes_called = True if self.list_nodes_empty: return None else: return_value = [type('node', (), {})()] return_value[0].id = self.node_id return_value[0].state = self.node_state return return_value def list_volumes(self, ex_resource_group=None): return_value = [type('volume', (), {})()] return_value[0].name = self.volume_name return return_value def ex_stop_node(self, node): self.ex_stop_node_called = True return self.stop def ex_start_node(self, node): self.ex_start_node_called = True return self.start def destroy_node(self, node=None, ex_destroy_vhd=False, ex_destroy_nic=False): self.destroy_node_called = True return self.terminate def destroy_volume(self, arg=None): self.destroy_volume_called = True return class TestProviderSpecific(TestCase): def __init__(self, *args, **kwargs): super(TestProviderSpecific, self).__init__(*args, **kwargs) self.command_arg = utils.default_args @unittest.mock.patch('os.system') @unittest.mock.patch('os.chmod') @unittest.mock.patch('os.path.isfile', side_effect=side_effect_create_local_rsa_key_pair_no_key) def test_create_local_rsa_key_pair_no_key(self, is_path_file, os_chmod, os_system): ssh = SSHCrossCloud(**self.command_arg) ssh.ssh_params.rsa_private_key_file_path = "a" ssh.spe_driver.create_local_rsa_key_pair() os_system.assert_called_with("ssh-keygen -b 2048 -f a") os_chmod.assert_called() is_path_file.assert_called() @unittest.mock.patch('os.system') @unittest.mock.patch('os.chmod') @unittest.mock.patch('os.path.isfile', side_effect=side_effect_create_local_rsa_key_pair_private) def test_create_local_rsa_key_pair_private_key(self, isfile, chmod, os_system): ssh = SSHCrossCloud(**self.command_arg) ssh.ssh_params.rsa_private_key_file_path = "a" ssh.spe_driver.create_local_rsa_key_pair() os_system.assert_called_with("ssh-keygen -b 2048 -y -f a > a.pub") @unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values') @unittest.mock.patch('os.path.isfile') def test_create_local_rsa_key_pair_private_key_and_public_key(self, isfile, update_custom_values): ssh = SSHCrossCloud(**self.command_arg) isfile.return_value = True ssh.ssh_params.rsa_private_key_file_path = "a" assert ssh.spe_driver.create_local_rsa_key_pair() == 0 def test_stop_instance_no_arg(self): ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "terminated" ssh.spe_driver.driver.stop = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.stop_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.ex_stop_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.stop = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.stop_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.ex_stop_node_called is True ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.stop = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.stop_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.ex_stop_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "terminated" ssh.spe_driver.driver.stop = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.stop_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.ex_stop_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.stop = False ssh.ssh_params.sshcrosscloud_instance_id = "a" with self.assertRaises(Exception): ssh.spe_driver.stop_instance_no_arg() def test_start_instance_no_arg(self): ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "not_stopped" ssh.spe_driver.driver.start = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.start_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.ex_start_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "stopped" ssh.spe_driver.driver.start = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.start_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.ex_start_node_called is True ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "stopped" ssh.spe_driver.driver.start = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.start_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.ex_start_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "not_stopped" ssh.spe_driver.driver.start = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.start_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.ex_start_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "stopped" ssh.spe_driver.driver.start = False ssh.ssh_params.sshcrosscloud_instance_id = "a" with self.assertRaises(Exception): ssh.spe_driver.start_instance_no_arg() def test_terminate_instance_no_arg(self): ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.terminate_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.destroy_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.terminate_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.destroy_node_called is True ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.terminate_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.destroy_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" assert ssh.spe_driver.terminate_instance_no_arg() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.destroy_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.terminate = False ssh.ssh_params.sshcrosscloud_instance_id = "a" with self.assertRaises(Exception): ssh.spe_driver.terminate_instance_no_arg() def test_get_node_any_arg(self): ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = 'foo' ssh.ssh_params.sshcrosscloud_instance_id = 'foo' assert ssh.spe_driver.get_node_any_arg().id == 'foo' ssh.spe_driver.driver.list_nodes_empty = True with self.assertRaises(Exception): ssh.spe_driver.get_node_any_arg() ssh.spe_driver.driver.node_id = 'foo' with self.assertRaises(Exception): ssh.spe_driver.get_node_any_arg() class TestSpecificAWS(TestCase): command_arg = utils.default_args @unittest.mock.patch('libcloud.compute.drivers.ec2.EC2NodeDriver') @unittest.mock.patch.object(sshcrosscloud.libcloud_extended.SpecificAWS, 'get_credentials') @unittest.mock.patch.object(sshcrosscloud.libcloud_extended.SpecificAWS, 'get_region_from_config_file') def test_init_specific(self, get_region, get_credentials, ec2_node_driver): os.environ['IMAGE_NAME'] = 'Fedora' ssh = SSHCrossCloud(**self.command_arg) ssh.ssh_params.aws.region = None get_credentials.return_value = "test1", "test2" ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params) assert ssh.spe_driver.init_specific() is not None assert ssh.ssh_params.instance_user == "fedora" get_region.assert_called() get_credentials.assert_called() ec2_node_driver.assert_called() @unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values') @unittest.mock.patch('os.path.isfile') @unittest.mock.patch('configparser.ConfigParser') def test_get_region(self, config_parser, is_file, update_custom_values): is_file.return_value = True ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params) assert ssh.spe_driver.get_region_from_config_file() is not None config_parser.assert_called() @unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values') @unittest.mock.patch('os.path.isfile') @unittest.mock.patch('configparser.ConfigParser') def test_get_credentials(self, config_parser, is_file, update_custom_values): is_file.return_value = True ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params) assert len(ssh.spe_driver.get_credentials()) is 2 config_parser.assert_called() class TestSpecificAzure(TestCase): def __init__(self, *args, **kwargs): super(TestSpecificAzure, self).__init__(*args, **kwargs) self.command_arg = utils.default_args self.command_arg['provider'] = 'azure' @unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values') @unittest.mock.patch('os.path.isfile') @unittest.mock.patch('configparser.ConfigParser') def test_get_credentials(self, config_parser, is_file, update_custom_values): is_file.return_value = True ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params) assert len(ssh.spe_driver.get_credentials()) is 4 config_parser.assert_called() def test_terminate_instance(self): ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" ssh.ssh_params.general_name = 'foo' ssh.spe_driver.driver.volume_name = 'foo' assert ssh.spe_driver.terminate_instance() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.destroy_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" ssh.ssh_params.general_name = 'foo' ssh.spe_driver.driver.volume_name = 'foo' assert ssh.spe_driver.terminate_instance() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.destroy_node_called is True ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" ssh.ssh_params.general_name = 'foo' ssh.spe_driver.driver.volume_name = 'foo' assert ssh.spe_driver.terminate_instance() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.destroy_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "b" ssh.spe_driver.driver.node_state = "terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" ssh.ssh_params.general_name = 'foo' ssh.spe_driver.driver.volume_name = 'foo' assert ssh.spe_driver.terminate_instance() is None assert ssh.spe_driver.driver.list_nodes_called is True assert ssh.spe_driver.driver.destroy_node_called is False ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.terminate = False ssh.ssh_params.sshcrosscloud_instance_id = "a" ssh.ssh_params.general_name = 'foo' ssh.spe_driver.driver.volume_name = 'foo' with self.assertRaises(Exception): ssh.spe_driver.terminate_instance() ssh.spe_driver.driver = FakeProviderDriver() ssh.spe_driver.driver.node_id = "a" ssh.spe_driver.driver.node_state = "not_terminated" ssh.spe_driver.driver.terminate = True ssh.ssh_params.sshcrosscloud_instance_id = "a" ssh.ssh_params.general_name = 'foo' ssh.spe_driver.driver.volume_name = 'foo' ssh.spe_driver.terminate_instance() assert ssh.spe_driver.driver.destroy_volume_called is True class TestSpecificGPC(TestCase): def __init__(self, *args, **kwargs): super(TestSpecificGPC, self).__init__(*args, **kwargs) self.command_arg = utils.default_args self.command_arg['provider'] = 'gcp' @unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values') @unittest.mock.patch('os.path.isfile') @unittest.mock.patch('configparser.ConfigParser') def test_get_credentials(self, config_parser, is_file, update_custom_values): is_file.return_value = True ssh = SSHCrossCloud(**self.command_arg) ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params) assert len(ssh.spe_driver.get_credentials()) is 4 config_parser.assert_called()
43.378049
102
0.705876
2,387
17,785
4.93297
0.056975
0.083567
0.167134
0.19414
0.85087
0.816645
0.80518
0.785902
0.771295
0.740892
0
0.001832
0.201912
17,785
409
103
43.484108
0.827744
0
0
0.694286
0
0
0.049873
0.009502
0
0
0
0
0.211429
1
0.071429
false
0
0.04
0
0.16
0
0
0
0
null
0
0
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
33e023b577adb0b77dfce41a990bd9ec1f875db1
4,965
py
Python
tests/hwsim/test_ap_pmf.py
yunus/Hostapd-with-WebID
2fb6f3a704629990af6dbd3a09d535788ba2d489
[ "Unlicense" ]
1
2021-11-12T10:52:06.000Z
2021-11-12T10:52:06.000Z
tests/hwsim/test_ap_pmf.py
arend/hostap
7c0e1e27575cd273f928fd4820047cba0322718c
[ "Unlicense" ]
null
null
null
tests/hwsim/test_ap_pmf.py
arend/hostap
7c0e1e27575cd273f928fd4820047cba0322718c
[ "Unlicense" ]
1
2021-11-12T10:52:06.000Z
2021-11-12T10:52:06.000Z
#!/usr/bin/python # # Protected management frames tests # Copyright (c) 2013, Jouni Malinen <j@w1.fi> # # This software may be distributed under the terms of the BSD license. # See README for more details. import time import subprocess import logging logger = logging.getLogger(__name__) import hwsim_utils import hostapd from wlantest import Wlantest def test_ap_pmf_required(dev, apdev): """WPA2-PSK AP with PMF required""" ssid = "test-pmf-required" wt = Wlantest() wt.flush() wt.add_passphrase("12345678") params = hostapd.wpa2_params(ssid=ssid, passphrase="12345678") params["wpa_key_mgmt"] = "WPA-PSK-SHA256"; params["ieee80211w"] = "2"; hostapd.add_ap(apdev[0]['ifname'], params) dev[0].connect(ssid, psk="12345678", ieee80211w="1", key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2") hwsim_utils.test_connectivity(dev[0].ifname, apdev[0]['ifname']) dev[1].connect(ssid, psk="12345678", ieee80211w="2", key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2") hwsim_utils.test_connectivity(dev[1].ifname, apdev[0]['ifname']) hapd = hostapd.Hostapd(apdev[0]['ifname']) hapd.request("SA_QUERY " + dev[0].p2p_interface_addr()) hapd.request("SA_QUERY " + dev[1].p2p_interface_addr()) wt.require_ap_pmf_mandatory(apdev[0]['bssid']) wt.require_sta_pmf(apdev[0]['bssid'], dev[0].p2p_interface_addr()) wt.require_sta_pmf_mandatory(apdev[0]['bssid'], dev[1].p2p_interface_addr()) time.sleep(0.1) if wt.get_sta_counter("valid_saqueryresp_tx", apdev[0]['bssid'], dev[0].p2p_interface_addr()) < 1: raise Exception("STA did not reply to SA Query") if wt.get_sta_counter("valid_saqueryresp_tx", apdev[0]['bssid'], dev[1].p2p_interface_addr()) < 1: raise Exception("STA did not reply to SA Query") def test_ap_pmf_optional(dev, apdev): """WPA2-PSK AP with PMF optional""" ssid = "test-pmf-optional" wt = Wlantest() wt.flush() wt.add_passphrase("12345678") params = hostapd.wpa2_params(ssid=ssid, passphrase="12345678") params["wpa_key_mgmt"] = "WPA-PSK"; params["ieee80211w"] = "1"; hostapd.add_ap(apdev[0]['ifname'], params) dev[0].connect(ssid, psk="12345678", ieee80211w="1", key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2") hwsim_utils.test_connectivity(dev[0].ifname, apdev[0]['ifname']) dev[1].connect(ssid, psk="12345678", ieee80211w="2", key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2") hwsim_utils.test_connectivity(dev[1].ifname, apdev[0]['ifname']) wt.require_ap_pmf_optional(apdev[0]['bssid']) wt.require_sta_pmf(apdev[0]['bssid'], dev[0].p2p_interface_addr()) wt.require_sta_pmf_mandatory(apdev[0]['bssid'], dev[1].p2p_interface_addr()) def test_ap_pmf_optional_2akm(dev, apdev): """WPA2-PSK AP with PMF optional (2 AKMs)""" ssid = "test-pmf-optional-2akm" wt = Wlantest() wt.flush() wt.add_passphrase("12345678") params = hostapd.wpa2_params(ssid=ssid, passphrase="12345678") params["wpa_key_mgmt"] = "WPA-PSK WPA-PSK-SHA256"; params["ieee80211w"] = "1"; hostapd.add_ap(apdev[0]['ifname'], params) dev[0].connect(ssid, psk="12345678", ieee80211w="1", key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2") hwsim_utils.test_connectivity(dev[0].ifname, apdev[0]['ifname']) dev[1].connect(ssid, psk="12345678", ieee80211w="2", key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2") hwsim_utils.test_connectivity(dev[1].ifname, apdev[0]['ifname']) wt.require_ap_pmf_optional(apdev[0]['bssid']) wt.require_sta_pmf(apdev[0]['bssid'], dev[0].p2p_interface_addr()) wt.require_sta_key_mgmt(apdev[0]['bssid'], dev[0].p2p_interface_addr(), "PSK-SHA256") wt.require_sta_pmf_mandatory(apdev[0]['bssid'], dev[1].p2p_interface_addr()) wt.require_sta_key_mgmt(apdev[0]['bssid'], dev[1].p2p_interface_addr(), "PSK-SHA256") def test_ap_pmf_negative(dev, apdev): """WPA2-PSK AP without PMF (negative test)""" ssid = "test-pmf-negative" wt = Wlantest() wt.flush() wt.add_passphrase("12345678") params = hostapd.wpa2_params(ssid=ssid, passphrase="12345678") hostapd.add_ap(apdev[0]['ifname'], params) dev[0].connect(ssid, psk="12345678", ieee80211w="1", key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2") hwsim_utils.test_connectivity(dev[0].ifname, apdev[0]['ifname']) try: dev[1].connect(ssid, psk="12345678", ieee80211w="2", key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2") hwsim_utils.test_connectivity(dev[1].ifname, apdev[0]['ifname']) raise Exception("PMF required STA connected to no PMF AP") except Exception, e: logger.debug("Ignore expected exception: " + str(e)) wt.require_ap_no_pmf(apdev[0]['bssid'])
44.72973
80
0.652971
701
4,965
4.443652
0.156919
0.052006
0.049438
0.045907
0.805778
0.752809
0.752809
0.731942
0.699839
0.699839
0
0.079872
0.180463
4,965
110
81
45.136364
0.685672
0.038671
0
0.645161
0
0
0.190797
0.004775
0
0
0
0
0
0
null
null
0.086022
0.064516
null
null
0
0
0
0
null
0
0
0
1
1
1
1
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
1
0
0
0
0
0
7
1d736cd5a6542dbd3c419f479be48f9d3af48e6d
10,140
py
Python
scrapepy/scrapepy.py
ivekmezovu/ScrapePy
f5ea2d82f7913c73b9ed9d35aa54e29768e50015
[ "MIT" ]
null
null
null
scrapepy/scrapepy.py
ivekmezovu/ScrapePy
f5ea2d82f7913c73b9ed9d35aa54e29768e50015
[ "MIT" ]
null
null
null
scrapepy/scrapepy.py
ivekmezovu/ScrapePy
f5ea2d82f7913c73b9ed9d35aa54e29768e50015
[ "MIT" ]
null
null
null
import urllib.request from urllib.request import urlretrieve import urllib from selenium import webdriver import os import time from selenium.common.exceptions import NoSuchElementException from tqdm import tqdm from .settings import Settings from .util import Util class ScrapePy: def __init__(self): self.options = webdriver.ChromeOptions() self.options.add_argument('headless') self.location = Settings.chromedriver_location self.driver = webdriver.Chrome(chrome_options=self.options, executable_path=self.location) def scrape_by_username(self, username): url = 'https://www.instagram.com/{}/'.format(username) driver = self.driver driver.get(url) if driver.find_elements_by_xpath("//div[@class='error-container']"): print(" >>> Wrong url!") elif Util.check_if_empty_profile(username) == True: print(" >>> Account doesn't have any post!") elif Util.check_if_private_profile(username) == True: print(" >>> Account is private!") else: try: srcs = [] for i in range(1, 7): driver.execute_script("window.scrollTo(0, document.body.scrollHeight);") time.sleep(2) img = driver.find_elements_by_tag_name('a') #'//img[@class="FFVAD"]')''' src = [s.get_attribute('href') for s in img if '.com/p/' in s.get_attribute('href')] [srcs.append(href) for href in src if href not in srcs] except NoSuchElementException: print(" >>> Can't find element!") src_col = [] i=0 for sr in srcs: i+=1 driver.get(sr) print("\n >>> Collecting {}/{}, -> {}".format(i, len(srcs), sr)) time.sleep(2) try: a = driver.find_element_by_xpath('//img[@class="FFVAD"]').get_attribute('src') print(" >>> Collected!") src_col.append(a) except NoSuchElementException: print(" >>> Can't find this element!") print(src_col) num = 1 for srces in src_col: try: '''for i in tqdm(range(len(src_col))): urllib.request.urlretrieve(srces, '{}.jpg'.format(num)) print(" >>> Downloaded {}/{})".format(num, len(src_col))) num+=1''' for f in tqdm(range(10), ascii=True, desc=" >>> Downloading"): pass urllib.request.urlretrieve(srces, '{}.jpg'.format(num)) print(" >>> Downloaded {}/{}".format(num, len(src_col))) num+=1 except Exception as e: time.sleep(2) print(" >>> Can't download, don't know, don't care") driver.close() def scrape_by_hashtag(self, hashtag): url = 'https://www.instagram.com/explore/tags/{}/'.format(hashtag) driver = self.driver driver.get(url) if driver.find_elements_by_xpath("//div[@class='error-container']"): print(" >>> Wrong url!") elif Util.check_if_empty_profile(username) == True: print(" >>> Account doesn't have any post!") elif Util.check_if_private_profile(username) == True: print(" >>> Account is private!") else: try: srcs = [] for i in range(1, 7): driver.execute_script("window.scrollTo(0, document.body.scrollHeight);") time.sleep(2) img = driver.find_elements_by_tag_name('a') #'//img[@class="FFVAD"]')''' src = [s.get_attribute('href') for s in img if '.com/p/' in s.get_attribute('href')] [srcs.append(href) for href in src if href not in srcs] except NoSuchElementException: print(" >>> Can't find element!") src_col = [] i=0 for sr in srcs: i+=1 driver.get(sr) print("\n >>> Collecting {}/{}, -> {}".format(i, len(srcs), sr)) time.sleep(2) try: a = driver.find_element_by_xpath('//img[@class="FFVAD"]').get_attribute('src') print(" >>> Collected!") src_col.append(a) except NoSuchElementException: print(" >>> Can't find this element!") print(src_col) num = 1 for srces in src_col: try: for f in tqdm(range(10), ascii=True, desc=" >>> Downloading"): pass urllib.request.urlretrieve(srces, '{}.jpg'.format(num)) print(" >>> Downloaded {}/{}".format(num, len(src_col))) num+=1 except Exception as e: time.sleep(2) print(" >>> Can't download, don't know, don't care") driver.close() def scrape_videos_by_username(self, username): url = 'https://www.instagram.com/{}/'.format(username) driver = self.driver driver.get(url) if driver.find_elements_by_xpath("//div[@class='error-container']"): print(" >>> Wrong url!") elif Util.check_if_empty_profile(username) == True: print(" >>> Account doesn't have any post!") elif Util.check_if_private_profile(username) == True: print(" >>> Account is private!") else: try: srcs = [] for i in range(1, 7): driver.execute_script("window.scrollTo(0, document.body.scrollHeight);") time.sleep(2) img = driver.find_elements_by_tag_name('a') #'//img[@class="FFVAD"]')''' src = [s.get_attribute('href') for s in img if '.com/p/' in s.get_attribute('href')] [srcs.append(href) for href in src if href not in srcs] except NoSuchElementException: print(" >>> Can't find element!") src_col = [] i=0 for sr in srcs: i+=1 driver.get(sr) print("\n >>> Collecting {}/{}, -> {}".format(i, len(srcs), sr)) time.sleep(2) try: a = driver.find_element_by_xpath('//video[@class="tWeCl"]').get_attribute('src') print(" >>> Collected!") src_col.append(a) except NoSuchElementException: print(" >>> Can't find this element!") print(src_col) num = 1 for srces in src_col: try: for f in tqdm(range(10), ascii=True, desc=" >>> Downloading"): pass urllib.request.urlretrieve(srces, '{}.mp4'.format(num)) print(" >>> Downloaded {}/{}".format(num, len(src_col))) num+=1 except Exception as e: time.sleep(2) print(" >>> Can't download, don't know, don't care") driver.close() def scrape_videos_by_hashtag(self, hashtag): url = 'https://www.instagram.com/explore/tags/{}/'.format(hashtag) driver = self.driver driver.get(url) if driver.find_elements_by_xpath("//div[@class='error-container']"): print(" >>> Wrong url!") elif Util.check_if_empty_profile(username) == True: print(" >>> Account doesn't have any post!") elif Util.check_if_private_profile(username) == True: print(" >>> Account is private!") else: try: srcs = [] for i in range(1, 7): driver.execute_script("window.scrollTo(0, document.body.scrollHeight);") time.sleep(2) img = driver.find_elements_by_tag_name('a') #'//img[@class="FFVAD"]')''' src = [s.get_attribute('href') for s in img if '.com/p/' in s.get_attribute('href')] [srcs.append(href) for href in src if href not in srcs] except NoSuchElementException: print(" >>> Can't find element!") src_col = [] i=0 for sr in srcs: i+=1 driver.get(sr) print("\n >>> Collecting {}/{}, -> {}".format(i, len(srcs), sr)) time.sleep(2) try: a = driver.find_element_by_xpath('//video[@class="tWeCl"]').get_attribute('src') print(" >>> Collected!") src_col.append(a) except NoSuchElementException: print(" >>> Can't find this element!") print(src_col) num = 1 for srces in src_col: try: for f in tqdm(range(10), ascii=True, desc=" >>> Downloading"): pass urllib.request.urlretrieve(srces, '{}.mp4'.format(num)) print(" >>> Downloaded {}/{}".format(num, len(src_col))) num+=1 except Exception as e: time.sleep(2) print(" >>> Can't download, don't know, don't care") driver.close()
31.786834
105
0.463905
1,037
10,140
4.420444
0.124397
0.028796
0.026178
0.019634
0.899651
0.899651
0.899651
0.899651
0.899651
0.899651
0
0.008493
0.407791
10,140
318
106
31.886792
0.754871
0.010651
0
0.901961
0
0
0.17539
0.034192
0
0
0
0
0
1
0.02451
false
0.019608
0.04902
0
0.078431
0.196078
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
d533837f54e91a91c940813975b0b7545b0b0cfd
68,666
py
Python
benchmarks/SimResults/_bigLittle_hrrs_spec_tugberk_pinned/cmp_GemsFDTD/power.py
TugberkArkose/MLScheduler
e493b6cbf7b9d29a2c9300d7dd6f0c2f102e4061
[ "Unlicense" ]
null
null
null
benchmarks/SimResults/_bigLittle_hrrs_spec_tugberk_pinned/cmp_GemsFDTD/power.py
TugberkArkose/MLScheduler
e493b6cbf7b9d29a2c9300d7dd6f0c2f102e4061
[ "Unlicense" ]
null
null
null
benchmarks/SimResults/_bigLittle_hrrs_spec_tugberk_pinned/cmp_GemsFDTD/power.py
TugberkArkose/MLScheduler
e493b6cbf7b9d29a2c9300d7dd6f0c2f102e4061
[ "Unlicense" ]
null
null
null
power = {'BUSES': {'Area': 1.33155, 'Bus/Area': 1.33155, 'Bus/Gate Leakage': 0.00662954, 'Bus/Peak Dynamic': 0.0, 'Bus/Runtime Dynamic': 0.0, 'Bus/Subthreshold Leakage': 0.0691322, 'Bus/Subthreshold Leakage with power gating': 0.0259246, 'Gate Leakage': 0.00662954, 'Peak Dynamic': 0.0, 'Runtime Dynamic': 0.0, 'Subthreshold Leakage': 0.0691322, 'Subthreshold Leakage with power gating': 0.0259246}, 'Core': [{'Area': 32.6082, 'Execution Unit/Area': 8.2042, 'Execution Unit/Complex ALUs/Area': 0.235435, 'Execution Unit/Complex ALUs/Gate Leakage': 0.0132646, 'Execution Unit/Complex ALUs/Peak Dynamic': 0.0704087, 'Execution Unit/Complex ALUs/Runtime Dynamic': 0.25799, 'Execution Unit/Complex ALUs/Subthreshold Leakage': 0.20111, 'Execution Unit/Complex ALUs/Subthreshold Leakage with power gating': 0.0754163, 'Execution Unit/Floating Point Units/Area': 4.6585, 'Execution Unit/Floating Point Units/Gate Leakage': 0.0656156, 'Execution Unit/Floating Point Units/Peak Dynamic': 0.397975, 'Execution Unit/Floating Point Units/Runtime Dynamic': 0.304033, 'Execution Unit/Floating Point Units/Subthreshold Leakage': 0.994829, 'Execution Unit/Floating Point Units/Subthreshold Leakage with power gating': 0.373061, 'Execution Unit/Gate Leakage': 0.122718, 'Execution Unit/Instruction Scheduler/Area': 2.17927, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Area': 0.328073, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Gate Leakage': 0.00115349, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Peak Dynamic': 1.20978, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Runtime Dynamic': 0.188804, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage': 0.017004, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage with power gating': 0.00962066, 'Execution Unit/Instruction Scheduler/Gate Leakage': 0.00730101, 'Execution Unit/Instruction Scheduler/Instruction Window/Area': 1.00996, 'Execution Unit/Instruction Scheduler/Instruction Window/Gate Leakage': 0.00529112, 'Execution Unit/Instruction Scheduler/Instruction Window/Peak Dynamic': 2.07911, 'Execution Unit/Instruction Scheduler/Instruction Window/Runtime Dynamic': 0.32694, 'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage': 0.0800117, 'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage with power gating': 0.0455351, 'Execution Unit/Instruction Scheduler/Peak Dynamic': 4.84781, 'Execution Unit/Instruction Scheduler/ROB/Area': 0.841232, 'Execution Unit/Instruction Scheduler/ROB/Gate Leakage': 0.000856399, 'Execution Unit/Instruction Scheduler/ROB/Peak Dynamic': 1.55892, 'Execution Unit/Instruction Scheduler/ROB/Runtime Dynamic': 0.187509, 'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage': 0.0178624, 'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage with power gating': 0.00897339, 'Execution Unit/Instruction Scheduler/Runtime Dynamic': 0.703254, 'Execution Unit/Instruction Scheduler/Subthreshold Leakage': 0.114878, 'Execution Unit/Instruction Scheduler/Subthreshold Leakage with power gating': 0.0641291, 'Execution Unit/Integer ALUs/Area': 0.47087, 'Execution Unit/Integer ALUs/Gate Leakage': 0.0265291, 'Execution Unit/Integer ALUs/Peak Dynamic': 0.125609, 'Execution Unit/Integer ALUs/Runtime Dynamic': 0.101344, 'Execution Unit/Integer ALUs/Subthreshold Leakage': 0.40222, 'Execution Unit/Integer ALUs/Subthreshold Leakage with power gating': 0.150833, 'Execution Unit/Peak Dynamic': 5.81006, 'Execution Unit/Register Files/Area': 0.570804, 'Execution Unit/Register Files/Floating Point RF/Area': 0.208131, 'Execution Unit/Register Files/Floating Point RF/Gate Leakage': 0.000232788, 'Execution Unit/Register Files/Floating Point RF/Peak Dynamic': 0.075186, 'Execution Unit/Register Files/Floating Point RF/Runtime Dynamic': 0.00684429, 'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage': 0.00399698, 'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage with power gating': 0.00176968, 'Execution Unit/Register Files/Gate Leakage': 0.000622708, 'Execution Unit/Register Files/Integer RF/Area': 0.362673, 'Execution Unit/Register Files/Integer RF/Gate Leakage': 0.00038992, 'Execution Unit/Register Files/Integer RF/Peak Dynamic': 0.075131, 'Execution Unit/Register Files/Integer RF/Runtime Dynamic': 0.0506177, 'Execution Unit/Register Files/Integer RF/Subthreshold Leakage': 0.00614175, 'Execution Unit/Register Files/Integer RF/Subthreshold Leakage with power gating': 0.00246675, 'Execution Unit/Register Files/Peak Dynamic': 0.150317, 'Execution Unit/Register Files/Runtime Dynamic': 0.057462, 'Execution Unit/Register Files/Subthreshold Leakage': 0.0101387, 'Execution Unit/Register Files/Subthreshold Leakage with power gating': 0.00423643, 'Execution Unit/Results Broadcast Bus/Area Overhead': 0.0442632, 'Execution Unit/Results Broadcast Bus/Gate Leakage': 0.00607074, 'Execution Unit/Results Broadcast Bus/Peak Dynamic': 0.200837, 'Execution Unit/Results Broadcast Bus/Runtime Dynamic': 0.549203, 'Execution Unit/Results Broadcast Bus/Subthreshold Leakage': 0.0920413, 'Execution Unit/Results Broadcast Bus/Subthreshold Leakage with power gating': 0.0345155, 'Execution Unit/Runtime Dynamic': 1.97329, 'Execution Unit/Subthreshold Leakage': 1.83518, 'Execution Unit/Subthreshold Leakage with power gating': 0.709678, 'Gate Leakage': 0.372997, 'Instruction Fetch Unit/Area': 5.86007, 'Instruction Fetch Unit/Branch Predictor/Area': 0.138516, 'Instruction Fetch Unit/Branch Predictor/Chooser/Area': 0.0435221, 'Instruction Fetch Unit/Branch Predictor/Chooser/Gate Leakage': 0.000278362, 'Instruction Fetch Unit/Branch Predictor/Chooser/Peak Dynamic': 0.0168831, 'Instruction Fetch Unit/Branch Predictor/Chooser/Runtime Dynamic': 0.000128293, 'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage': 0.00759719, 'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage with power gating': 0.0039236, 'Instruction Fetch Unit/Branch Predictor/Gate Leakage': 0.000757657, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Area': 0.0435221, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Gate Leakage': 0.000278362, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Peak Dynamic': 0.0168831, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Runtime Dynamic': 0.000128293, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage': 0.00759719, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage with power gating': 0.0039236, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Area': 0.0257064, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Gate Leakage': 0.000154548, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Peak Dynamic': 0.0142575, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Runtime Dynamic': 0.000111033, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage': 0.00384344, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage with power gating': 0.00198631, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Area': 0.0151917, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Gate Leakage': 8.00196e-05, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Peak Dynamic': 0.00527447, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Runtime Dynamic': 4.25945e-05, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage': 0.00181347, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage with power gating': 0.000957045, 'Instruction Fetch Unit/Branch Predictor/Peak Dynamic': 0.0597838, 'Instruction Fetch Unit/Branch Predictor/RAS/Area': 0.0105732, 'Instruction Fetch Unit/Branch Predictor/RAS/Gate Leakage': 4.63858e-05, 'Instruction Fetch Unit/Branch Predictor/RAS/Peak Dynamic': 0.0117602, 'Instruction Fetch Unit/Branch Predictor/RAS/Runtime Dynamic': 0.000727128, 'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage': 0.000932505, 'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage with power gating': 0.000494733, 'Instruction Fetch Unit/Branch Predictor/Runtime Dynamic': 0.00109475, 'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage': 0.0199703, 'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage with power gating': 0.0103282, 'Instruction Fetch Unit/Branch Target Buffer/Area': 0.64954, 'Instruction Fetch Unit/Branch Target Buffer/Gate Leakage': 0.00272758, 'Instruction Fetch Unit/Branch Target Buffer/Peak Dynamic': 0.177867, 'Instruction Fetch Unit/Branch Target Buffer/Runtime Dynamic': 0.00125542, 'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage': 0.0811682, 'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage with power gating': 0.0435357, 'Instruction Fetch Unit/Gate Leakage': 0.0590479, 'Instruction Fetch Unit/Instruction Buffer/Area': 0.0226323, 'Instruction Fetch Unit/Instruction Buffer/Gate Leakage': 6.83558e-05, 'Instruction Fetch Unit/Instruction Buffer/Peak Dynamic': 0.606827, 'Instruction Fetch Unit/Instruction Buffer/Runtime Dynamic': 0.0486601, 'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage': 0.00151885, 'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage with power gating': 0.000701682, 'Instruction Fetch Unit/Instruction Cache/Area': 3.14635, 'Instruction Fetch Unit/Instruction Cache/Gate Leakage': 0.029931, 'Instruction Fetch Unit/Instruction Cache/Peak Dynamic': 3.0952, 'Instruction Fetch Unit/Instruction Cache/Runtime Dynamic': 0.117729, 'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage': 0.367022, 'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage with power gating': 0.180386, 'Instruction Fetch Unit/Instruction Decoder/Area': 1.85799, 'Instruction Fetch Unit/Instruction Decoder/Gate Leakage': 0.0222493, 'Instruction Fetch Unit/Instruction Decoder/Peak Dynamic': 1.37404, 'Instruction Fetch Unit/Instruction Decoder/Runtime Dynamic': 0.165272, 'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage': 0.442943, 'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage with power gating': 0.166104, 'Instruction Fetch Unit/Peak Dynamic': 5.46623, 'Instruction Fetch Unit/Runtime Dynamic': 0.334011, 'Instruction Fetch Unit/Subthreshold Leakage': 0.932587, 'Instruction Fetch Unit/Subthreshold Leakage with power gating': 0.408542, 'L2/Area': 4.53318, 'L2/Gate Leakage': 0.015464, 'L2/Peak Dynamic': 0.130932, 'L2/Runtime Dynamic': 0.037829, 'L2/Subthreshold Leakage': 0.834142, 'L2/Subthreshold Leakage with power gating': 0.401066, 'Load Store Unit/Area': 8.80969, 'Load Store Unit/Data Cache/Area': 6.84535, 'Load Store Unit/Data Cache/Gate Leakage': 0.0279261, 'Load Store Unit/Data Cache/Peak Dynamic': 3.29559, 'Load Store Unit/Data Cache/Runtime Dynamic': 1.04058, 'Load Store Unit/Data Cache/Subthreshold Leakage': 0.527675, 'Load Store Unit/Data Cache/Subthreshold Leakage with power gating': 0.25085, 'Load Store Unit/Gate Leakage': 0.0351387, 'Load Store Unit/LoadQ/Area': 0.0836782, 'Load Store Unit/LoadQ/Gate Leakage': 0.00059896, 'Load Store Unit/LoadQ/Peak Dynamic': 0.0665964, 'Load Store Unit/LoadQ/Runtime Dynamic': 0.0665963, 'Load Store Unit/LoadQ/Subthreshold Leakage': 0.00941961, 'Load Store Unit/LoadQ/Subthreshold Leakage with power gating': 0.00536918, 'Load Store Unit/Peak Dynamic': 3.61135, 'Load Store Unit/Runtime Dynamic': 1.4356, 'Load Store Unit/StoreQ/Area': 0.322079, 'Load Store Unit/StoreQ/Gate Leakage': 0.00329971, 'Load Store Unit/StoreQ/Peak Dynamic': 0.164215, 'Load Store Unit/StoreQ/Runtime Dynamic': 0.32843, 'Load Store Unit/StoreQ/Subthreshold Leakage': 0.0345621, 'Load Store Unit/StoreQ/Subthreshold Leakage with power gating': 0.0197004, 'Load Store Unit/Subthreshold Leakage': 0.591622, 'Load Store Unit/Subthreshold Leakage with power gating': 0.283406, 'Memory Management Unit/Area': 0.434579, 'Memory Management Unit/Dtlb/Area': 0.0879726, 'Memory Management Unit/Dtlb/Gate Leakage': 0.00088729, 'Memory Management Unit/Dtlb/Peak Dynamic': 0.0582806, 'Memory Management Unit/Dtlb/Runtime Dynamic': 0.0602397, 'Memory Management Unit/Dtlb/Subthreshold Leakage': 0.0155699, 'Memory Management Unit/Dtlb/Subthreshold Leakage with power gating': 0.00887485, 'Memory Management Unit/Gate Leakage': 0.00813591, 'Memory Management Unit/Itlb/Area': 0.301552, 'Memory Management Unit/Itlb/Gate Leakage': 0.00393464, 'Memory Management Unit/Itlb/Peak Dynamic': 0.192448, 'Memory Management Unit/Itlb/Runtime Dynamic': 0.0193218, 'Memory Management Unit/Itlb/Subthreshold Leakage': 0.0413758, 'Memory Management Unit/Itlb/Subthreshold Leakage with power gating': 0.0235842, 'Memory Management Unit/Peak Dynamic': 0.451703, 'Memory Management Unit/Runtime Dynamic': 0.0795615, 'Memory Management Unit/Subthreshold Leakage': 0.0769113, 'Memory Management Unit/Subthreshold Leakage with power gating': 0.0399462, 'Peak Dynamic': 20.032, 'Renaming Unit/Area': 0.369768, 'Renaming Unit/FP Front End RAT/Area': 0.168486, 'Renaming Unit/FP Front End RAT/Gate Leakage': 0.00489731, 'Renaming Unit/FP Front End RAT/Peak Dynamic': 3.33511, 'Renaming Unit/FP Front End RAT/Runtime Dynamic': 0.262307, 'Renaming Unit/FP Front End RAT/Subthreshold Leakage': 0.0437281, 'Renaming Unit/FP Front End RAT/Subthreshold Leakage with power gating': 0.024925, 'Renaming Unit/Free List/Area': 0.0414755, 'Renaming Unit/Free List/Gate Leakage': 4.15911e-05, 'Renaming Unit/Free List/Peak Dynamic': 0.0401324, 'Renaming Unit/Free List/Runtime Dynamic': 0.0128108, 'Renaming Unit/Free List/Subthreshold Leakage': 0.000670426, 'Renaming Unit/Free List/Subthreshold Leakage with power gating': 0.000377987, 'Renaming Unit/Gate Leakage': 0.00863632, 'Renaming Unit/Int Front End RAT/Area': 0.114751, 'Renaming Unit/Int Front End RAT/Gate Leakage': 0.00038343, 'Renaming Unit/Int Front End RAT/Peak Dynamic': 0.86945, 'Renaming Unit/Int Front End RAT/Runtime Dynamic': 0.09465, 'Renaming Unit/Int Front End RAT/Subthreshold Leakage': 0.00611897, 'Renaming Unit/Int Front End RAT/Subthreshold Leakage with power gating': 0.00348781, 'Renaming Unit/Peak Dynamic': 4.56169, 'Renaming Unit/Runtime Dynamic': 0.369768, 'Renaming Unit/Subthreshold Leakage': 0.070483, 'Renaming Unit/Subthreshold Leakage with power gating': 0.0362779, 'Runtime Dynamic': 4.23006, 'Subthreshold Leakage': 6.21877, 'Subthreshold Leakage with power gating': 2.58311}, {'Area': 32.0201, 'Execution Unit/Area': 7.68434, 'Execution Unit/Complex ALUs/Area': 0.235435, 'Execution Unit/Complex ALUs/Gate Leakage': 0.0132646, 'Execution Unit/Complex ALUs/Peak Dynamic': 0.023245, 'Execution Unit/Complex ALUs/Runtime Dynamic': 0.220946, 'Execution Unit/Complex ALUs/Subthreshold Leakage': 0.20111, 'Execution Unit/Complex ALUs/Subthreshold Leakage with power gating': 0.0754163, 'Execution Unit/Floating Point Units/Area': 4.6585, 'Execution Unit/Floating Point Units/Gate Leakage': 0.0656156, 'Execution Unit/Floating Point Units/Peak Dynamic': 0.131656, 'Execution Unit/Floating Point Units/Runtime Dynamic': 0.304033, 'Execution Unit/Floating Point Units/Subthreshold Leakage': 0.994829, 'Execution Unit/Floating Point Units/Subthreshold Leakage with power gating': 0.373061, 'Execution Unit/Gate Leakage': 0.120359, 'Execution Unit/Instruction Scheduler/Area': 1.66526, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Area': 0.275653, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Gate Leakage': 0.000977433, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Peak Dynamic': 1.04181, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Runtime Dynamic': 0.0536798, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage': 0.0143453, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage with power gating': 0.00810519, 'Execution Unit/Instruction Scheduler/Gate Leakage': 0.00568913, 'Execution Unit/Instruction Scheduler/Instruction Window/Area': 0.805223, 'Execution Unit/Instruction Scheduler/Instruction Window/Gate Leakage': 0.00414562, 'Execution Unit/Instruction Scheduler/Instruction Window/Peak Dynamic': 1.6763, 'Execution Unit/Instruction Scheduler/Instruction Window/Runtime Dynamic': 0.0865835, 'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage': 0.0625755, 'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage with power gating': 0.0355964, 'Execution Unit/Instruction Scheduler/Peak Dynamic': 3.82262, 'Execution Unit/Instruction Scheduler/ROB/Area': 0.584388, 'Execution Unit/Instruction Scheduler/ROB/Gate Leakage': 0.00056608, 'Execution Unit/Instruction Scheduler/ROB/Peak Dynamic': 1.10451, 'Execution Unit/Instruction Scheduler/ROB/Runtime Dynamic': 0.0437044, 'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage': 0.00906853, 'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage with power gating': 0.00364446, 'Execution Unit/Instruction Scheduler/Runtime Dynamic': 0.183968, 'Execution Unit/Instruction Scheduler/Subthreshold Leakage': 0.0859892, 'Execution Unit/Instruction Scheduler/Subthreshold Leakage with power gating': 0.047346, 'Execution Unit/Integer ALUs/Area': 0.47087, 'Execution Unit/Integer ALUs/Gate Leakage': 0.0265291, 'Execution Unit/Integer ALUs/Peak Dynamic': 0.0412103, 'Execution Unit/Integer ALUs/Runtime Dynamic': 0.101344, 'Execution Unit/Integer ALUs/Subthreshold Leakage': 0.40222, 'Execution Unit/Integer ALUs/Subthreshold Leakage with power gating': 0.150833, 'Execution Unit/Peak Dynamic': 4.13179, 'Execution Unit/Register Files/Area': 0.570804, 'Execution Unit/Register Files/Floating Point RF/Area': 0.208131, 'Execution Unit/Register Files/Floating Point RF/Gate Leakage': 0.000232788, 'Execution Unit/Register Files/Floating Point RF/Peak Dynamic': 0.0248725, 'Execution Unit/Register Files/Floating Point RF/Runtime Dynamic': 0.00225157, 'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage': 0.00399698, 'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage with power gating': 0.00176968, 'Execution Unit/Register Files/Gate Leakage': 0.000622708, 'Execution Unit/Register Files/Integer RF/Area': 0.362673, 'Execution Unit/Register Files/Integer RF/Gate Leakage': 0.00038992, 'Execution Unit/Register Files/Integer RF/Peak Dynamic': 0.0247354, 'Execution Unit/Register Files/Integer RF/Runtime Dynamic': 0.0166518, 'Execution Unit/Register Files/Integer RF/Subthreshold Leakage': 0.00614175, 'Execution Unit/Register Files/Integer RF/Subthreshold Leakage with power gating': 0.00246675, 'Execution Unit/Register Files/Peak Dynamic': 0.049608, 'Execution Unit/Register Files/Runtime Dynamic': 0.0189033, 'Execution Unit/Register Files/Subthreshold Leakage': 0.0101387, 'Execution Unit/Register Files/Subthreshold Leakage with power gating': 0.00423643, 'Execution Unit/Results Broadcast Bus/Area Overhead': 0.0390912, 'Execution Unit/Results Broadcast Bus/Gate Leakage': 0.00537402, 'Execution Unit/Results Broadcast Bus/Peak Dynamic': 0.0579232, 'Execution Unit/Results Broadcast Bus/Runtime Dynamic': 0.15908, 'Execution Unit/Results Broadcast Bus/Subthreshold Leakage': 0.081478, 'Execution Unit/Results Broadcast Bus/Subthreshold Leakage with power gating': 0.0305543, 'Execution Unit/Runtime Dynamic': 0.988274, 'Execution Unit/Subthreshold Leakage': 1.79543, 'Execution Unit/Subthreshold Leakage with power gating': 0.688821, 'Gate Leakage': 0.368936, 'Instruction Fetch Unit/Area': 5.85939, 'Instruction Fetch Unit/Branch Predictor/Area': 0.138516, 'Instruction Fetch Unit/Branch Predictor/Chooser/Area': 0.0435221, 'Instruction Fetch Unit/Branch Predictor/Chooser/Gate Leakage': 0.000278362, 'Instruction Fetch Unit/Branch Predictor/Chooser/Peak Dynamic': 0.0168831, 'Instruction Fetch Unit/Branch Predictor/Chooser/Runtime Dynamic': 4.39474e-05, 'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage': 0.00759719, 'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage with power gating': 0.0039236, 'Instruction Fetch Unit/Branch Predictor/Gate Leakage': 0.000757657, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Area': 0.0435221, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Gate Leakage': 0.000278362, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Peak Dynamic': 0.0168831, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Runtime Dynamic': 4.39474e-05, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage': 0.00759719, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage with power gating': 0.0039236, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Area': 0.0257064, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Gate Leakage': 0.000154548, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Peak Dynamic': 0.0142575, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Runtime Dynamic': 3.84241e-05, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage': 0.00384344, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage with power gating': 0.00198631, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Area': 0.0151917, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Gate Leakage': 8.00196e-05, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Peak Dynamic': 0.00527447, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Runtime Dynamic': 1.49544e-05, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage': 0.00181347, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage with power gating': 0.000957045, 'Instruction Fetch Unit/Branch Predictor/Peak Dynamic': 0.0597838, 'Instruction Fetch Unit/Branch Predictor/RAS/Area': 0.0105732, 'Instruction Fetch Unit/Branch Predictor/RAS/Gate Leakage': 4.63858e-05, 'Instruction Fetch Unit/Branch Predictor/RAS/Peak Dynamic': 0.0117602, 'Instruction Fetch Unit/Branch Predictor/RAS/Runtime Dynamic': 0.000239204, 'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage': 0.000932505, 'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage with power gating': 0.000494733, 'Instruction Fetch Unit/Branch Predictor/Runtime Dynamic': 0.000365523, 'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage': 0.0199703, 'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage with power gating': 0.0103282, 'Instruction Fetch Unit/Branch Target Buffer/Area': 0.64954, 'Instruction Fetch Unit/Branch Target Buffer/Gate Leakage': 0.00272758, 'Instruction Fetch Unit/Branch Target Buffer/Peak Dynamic': 0.177867, 'Instruction Fetch Unit/Branch Target Buffer/Runtime Dynamic': 0.000416149, 'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage': 0.0811682, 'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage with power gating': 0.0435357, 'Instruction Fetch Unit/Gate Leakage': 0.0589979, 'Instruction Fetch Unit/Instruction Buffer/Area': 0.0226323, 'Instruction Fetch Unit/Instruction Buffer/Gate Leakage': 6.83558e-05, 'Instruction Fetch Unit/Instruction Buffer/Peak Dynamic': 0.606827, 'Instruction Fetch Unit/Instruction Buffer/Runtime Dynamic': 0.0160078, 'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage': 0.00151885, 'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage with power gating': 0.000701682, 'Instruction Fetch Unit/Instruction Cache/Area': 3.14635, 'Instruction Fetch Unit/Instruction Cache/Gate Leakage': 0.029931, 'Instruction Fetch Unit/Instruction Cache/Peak Dynamic': 1.01823, 'Instruction Fetch Unit/Instruction Cache/Runtime Dynamic': 0.038844, 'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage': 0.367022, 'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage with power gating': 0.180386, 'Instruction Fetch Unit/Instruction Decoder/Area': 1.85799, 'Instruction Fetch Unit/Instruction Decoder/Gate Leakage': 0.0222493, 'Instruction Fetch Unit/Instruction Decoder/Peak Dynamic': 1.37404, 'Instruction Fetch Unit/Instruction Decoder/Runtime Dynamic': 0.0543695, 'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage': 0.442943, 'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage with power gating': 0.166104, 'Instruction Fetch Unit/Peak Dynamic': 3.28616, 'Instruction Fetch Unit/Runtime Dynamic': 0.110003, 'Instruction Fetch Unit/Subthreshold Leakage': 0.932286, 'Instruction Fetch Unit/Subthreshold Leakage with power gating': 0.40843, 'L2/Area': 4.53318, 'L2/Gate Leakage': 0.015464, 'L2/Peak Dynamic': 0.0427196, 'L2/Runtime Dynamic': 0.0124854, 'L2/Subthreshold Leakage': 0.834142, 'L2/Subthreshold Leakage with power gating': 0.401066, 'Load Store Unit/Area': 8.80901, 'Load Store Unit/Data Cache/Area': 6.84535, 'Load Store Unit/Data Cache/Gate Leakage': 0.0279261, 'Load Store Unit/Data Cache/Peak Dynamic': 1.9148, 'Load Store Unit/Data Cache/Runtime Dynamic': 0.342833, 'Load Store Unit/Data Cache/Subthreshold Leakage': 0.527675, 'Load Store Unit/Data Cache/Subthreshold Leakage with power gating': 0.25085, 'Load Store Unit/Gate Leakage': 0.0350888, 'Load Store Unit/LoadQ/Area': 0.0836782, 'Load Store Unit/LoadQ/Gate Leakage': 0.00059896, 'Load Store Unit/LoadQ/Peak Dynamic': 0.0219246, 'Load Store Unit/LoadQ/Runtime Dynamic': 0.0219246, 'Load Store Unit/LoadQ/Subthreshold Leakage': 0.00941961, 'Load Store Unit/LoadQ/Subthreshold Leakage with power gating': 0.00536918, 'Load Store Unit/Peak Dynamic': 2.01833, 'Load Store Unit/Runtime Dynamic': 0.472882, 'Load Store Unit/StoreQ/Area': 0.322079, 'Load Store Unit/StoreQ/Gate Leakage': 0.00329971, 'Load Store Unit/StoreQ/Peak Dynamic': 0.0540624, 'Load Store Unit/StoreQ/Runtime Dynamic': 0.108125, 'Load Store Unit/StoreQ/Subthreshold Leakage': 0.0345621, 'Load Store Unit/StoreQ/Subthreshold Leakage with power gating': 0.0197004, 'Load Store Unit/Subthreshold Leakage': 0.591321, 'Load Store Unit/Subthreshold Leakage with power gating': 0.283293, 'Memory Management Unit/Area': 0.4339, 'Memory Management Unit/Dtlb/Area': 0.0879726, 'Memory Management Unit/Dtlb/Gate Leakage': 0.00088729, 'Memory Management Unit/Dtlb/Peak Dynamic': 0.0191869, 'Memory Management Unit/Dtlb/Runtime Dynamic': 0.0198261, 'Memory Management Unit/Dtlb/Subthreshold Leakage': 0.0155699, 'Memory Management Unit/Dtlb/Subthreshold Leakage with power gating': 0.00887485, 'Memory Management Unit/Gate Leakage': 0.00808595, 'Memory Management Unit/Itlb/Area': 0.301552, 'Memory Management Unit/Itlb/Gate Leakage': 0.00393464, 'Memory Management Unit/Itlb/Peak Dynamic': 0.06331, 'Memory Management Unit/Itlb/Runtime Dynamic': 0.0063751, 'Memory Management Unit/Itlb/Subthreshold Leakage': 0.0413758, 'Memory Management Unit/Itlb/Subthreshold Leakage with power gating': 0.0235842, 'Memory Management Unit/Peak Dynamic': 0.252379, 'Memory Management Unit/Runtime Dynamic': 0.0262012, 'Memory Management Unit/Subthreshold Leakage': 0.0766103, 'Memory Management Unit/Subthreshold Leakage with power gating': 0.0398333, 'Peak Dynamic': 13.3209, 'Renaming Unit/Area': 0.303608, 'Renaming Unit/FP Front End RAT/Area': 0.131045, 'Renaming Unit/FP Front End RAT/Gate Leakage': 0.00351123, 'Renaming Unit/FP Front End RAT/Peak Dynamic': 2.51468, 'Renaming Unit/FP Front End RAT/Runtime Dynamic': 0.0654283, 'Renaming Unit/FP Front End RAT/Subthreshold Leakage': 0.0308571, 'Renaming Unit/FP Front End RAT/Subthreshold Leakage with power gating': 0.0175885, 'Renaming Unit/Free List/Area': 0.0340654, 'Renaming Unit/Free List/Gate Leakage': 2.5481e-05, 'Renaming Unit/Free List/Peak Dynamic': 0.0306032, 'Renaming Unit/Free List/Runtime Dynamic': 0.00321813, 'Renaming Unit/Free List/Subthreshold Leakage': 0.000370144, 'Renaming Unit/Free List/Subthreshold Leakage with power gating': 0.000201064, 'Renaming Unit/Gate Leakage': 0.00708398, 'Renaming Unit/Int Front End RAT/Area': 0.0941223, 'Renaming Unit/Int Front End RAT/Gate Leakage': 0.000283242, 'Renaming Unit/Int Front End RAT/Peak Dynamic': 0.731965, 'Renaming Unit/Int Front End RAT/Runtime Dynamic': 0.0264341, 'Renaming Unit/Int Front End RAT/Subthreshold Leakage': 0.00435488, 'Renaming Unit/Int Front End RAT/Subthreshold Leakage with power gating': 0.00248228, 'Renaming Unit/Peak Dynamic': 3.58947, 'Renaming Unit/Runtime Dynamic': 0.0950806, 'Renaming Unit/Subthreshold Leakage': 0.0552466, 'Renaming Unit/Subthreshold Leakage with power gating': 0.0276461, 'Runtime Dynamic': 1.70493, 'Subthreshold Leakage': 6.16288, 'Subthreshold Leakage with power gating': 2.55328}, {'Area': 32.0201, 'Execution Unit/Area': 7.68434, 'Execution Unit/Complex ALUs/Area': 0.235435, 'Execution Unit/Complex ALUs/Gate Leakage': 0.0132646, 'Execution Unit/Complex ALUs/Peak Dynamic': 0.0232413, 'Execution Unit/Complex ALUs/Runtime Dynamic': 0.220943, 'Execution Unit/Complex ALUs/Subthreshold Leakage': 0.20111, 'Execution Unit/Complex ALUs/Subthreshold Leakage with power gating': 0.0754163, 'Execution Unit/Floating Point Units/Area': 4.6585, 'Execution Unit/Floating Point Units/Gate Leakage': 0.0656156, 'Execution Unit/Floating Point Units/Peak Dynamic': 0.13164, 'Execution Unit/Floating Point Units/Runtime Dynamic': 0.304033, 'Execution Unit/Floating Point Units/Subthreshold Leakage': 0.994829, 'Execution Unit/Floating Point Units/Subthreshold Leakage with power gating': 0.373061, 'Execution Unit/Gate Leakage': 0.120359, 'Execution Unit/Instruction Scheduler/Area': 1.66526, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Area': 0.275653, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Gate Leakage': 0.000977433, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Peak Dynamic': 1.04181, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Runtime Dynamic': 0.0536686, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage': 0.0143453, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage with power gating': 0.00810519, 'Execution Unit/Instruction Scheduler/Gate Leakage': 0.00568913, 'Execution Unit/Instruction Scheduler/Instruction Window/Area': 0.805223, 'Execution Unit/Instruction Scheduler/Instruction Window/Gate Leakage': 0.00414562, 'Execution Unit/Instruction Scheduler/Instruction Window/Peak Dynamic': 1.6763, 'Execution Unit/Instruction Scheduler/Instruction Window/Runtime Dynamic': 0.0865655, 'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage': 0.0625755, 'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage with power gating': 0.0355964, 'Execution Unit/Instruction Scheduler/Peak Dynamic': 3.82262, 'Execution Unit/Instruction Scheduler/ROB/Area': 0.584388, 'Execution Unit/Instruction Scheduler/ROB/Gate Leakage': 0.00056608, 'Execution Unit/Instruction Scheduler/ROB/Peak Dynamic': 1.10451, 'Execution Unit/Instruction Scheduler/ROB/Runtime Dynamic': 0.0436953, 'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage': 0.00906853, 'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage with power gating': 0.00364446, 'Execution Unit/Instruction Scheduler/Runtime Dynamic': 0.183929, 'Execution Unit/Instruction Scheduler/Subthreshold Leakage': 0.0859892, 'Execution Unit/Instruction Scheduler/Subthreshold Leakage with power gating': 0.047346, 'Execution Unit/Integer ALUs/Area': 0.47087, 'Execution Unit/Integer ALUs/Gate Leakage': 0.0265291, 'Execution Unit/Integer ALUs/Peak Dynamic': 0.0411987, 'Execution Unit/Integer ALUs/Runtime Dynamic': 0.101344, 'Execution Unit/Integer ALUs/Subthreshold Leakage': 0.40222, 'Execution Unit/Integer ALUs/Subthreshold Leakage with power gating': 0.150833, 'Execution Unit/Peak Dynamic': 4.13174, 'Execution Unit/Register Files/Area': 0.570804, 'Execution Unit/Register Files/Floating Point RF/Area': 0.208131, 'Execution Unit/Register Files/Floating Point RF/Gate Leakage': 0.000232788, 'Execution Unit/Register Files/Floating Point RF/Peak Dynamic': 0.0248697, 'Execution Unit/Register Files/Floating Point RF/Runtime Dynamic': 0.0022511, 'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage': 0.00399698, 'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage with power gating': 0.00176968, 'Execution Unit/Register Files/Gate Leakage': 0.000622708, 'Execution Unit/Register Files/Integer RF/Area': 0.362673, 'Execution Unit/Register Files/Integer RF/Gate Leakage': 0.00038992, 'Execution Unit/Register Files/Integer RF/Peak Dynamic': 0.0247301, 'Execution Unit/Register Files/Integer RF/Runtime Dynamic': 0.0166483, 'Execution Unit/Register Files/Integer RF/Subthreshold Leakage': 0.00614175, 'Execution Unit/Register Files/Integer RF/Subthreshold Leakage with power gating': 0.00246675, 'Execution Unit/Register Files/Peak Dynamic': 0.0495998, 'Execution Unit/Register Files/Runtime Dynamic': 0.0188994, 'Execution Unit/Register Files/Subthreshold Leakage': 0.0101387, 'Execution Unit/Register Files/Subthreshold Leakage with power gating': 0.00423643, 'Execution Unit/Results Broadcast Bus/Area Overhead': 0.0390912, 'Execution Unit/Results Broadcast Bus/Gate Leakage': 0.00537402, 'Execution Unit/Results Broadcast Bus/Peak Dynamic': 0.057911, 'Execution Unit/Results Broadcast Bus/Runtime Dynamic': 0.15905, 'Execution Unit/Results Broadcast Bus/Subthreshold Leakage': 0.081478, 'Execution Unit/Results Broadcast Bus/Subthreshold Leakage with power gating': 0.0305543, 'Execution Unit/Runtime Dynamic': 0.9882, 'Execution Unit/Subthreshold Leakage': 1.79543, 'Execution Unit/Subthreshold Leakage with power gating': 0.688821, 'Gate Leakage': 0.368936, 'Instruction Fetch Unit/Area': 5.85939, 'Instruction Fetch Unit/Branch Predictor/Area': 0.138516, 'Instruction Fetch Unit/Branch Predictor/Chooser/Area': 0.0435221, 'Instruction Fetch Unit/Branch Predictor/Chooser/Gate Leakage': 0.000278362, 'Instruction Fetch Unit/Branch Predictor/Chooser/Peak Dynamic': 0.0168831, 'Instruction Fetch Unit/Branch Predictor/Chooser/Runtime Dynamic': 4.39436e-05, 'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage': 0.00759719, 'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage with power gating': 0.0039236, 'Instruction Fetch Unit/Branch Predictor/Gate Leakage': 0.000757657, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Area': 0.0435221, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Gate Leakage': 0.000278362, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Peak Dynamic': 0.0168831, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Runtime Dynamic': 4.39436e-05, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage': 0.00759719, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage with power gating': 0.0039236, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Area': 0.0257064, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Gate Leakage': 0.000154548, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Peak Dynamic': 0.0142575, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Runtime Dynamic': 3.84208e-05, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage': 0.00384344, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage with power gating': 0.00198631, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Area': 0.0151917, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Gate Leakage': 8.00196e-05, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Peak Dynamic': 0.00527447, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Runtime Dynamic': 1.49532e-05, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage': 0.00181347, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage with power gating': 0.000957045, 'Instruction Fetch Unit/Branch Predictor/Peak Dynamic': 0.0597838, 'Instruction Fetch Unit/Branch Predictor/RAS/Area': 0.0105732, 'Instruction Fetch Unit/Branch Predictor/RAS/Gate Leakage': 4.63858e-05, 'Instruction Fetch Unit/Branch Predictor/RAS/Peak Dynamic': 0.0117602, 'Instruction Fetch Unit/Branch Predictor/RAS/Runtime Dynamic': 0.000239154, 'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage': 0.000932505, 'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage with power gating': 0.000494733, 'Instruction Fetch Unit/Branch Predictor/Runtime Dynamic': 0.000365462, 'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage': 0.0199703, 'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage with power gating': 0.0103282, 'Instruction Fetch Unit/Branch Target Buffer/Area': 0.64954, 'Instruction Fetch Unit/Branch Target Buffer/Gate Leakage': 0.00272758, 'Instruction Fetch Unit/Branch Target Buffer/Peak Dynamic': 0.177867, 'Instruction Fetch Unit/Branch Target Buffer/Runtime Dynamic': 0.000416112, 'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage': 0.0811682, 'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage with power gating': 0.0435357, 'Instruction Fetch Unit/Gate Leakage': 0.0589979, 'Instruction Fetch Unit/Instruction Buffer/Area': 0.0226323, 'Instruction Fetch Unit/Instruction Buffer/Gate Leakage': 6.83558e-05, 'Instruction Fetch Unit/Instruction Buffer/Peak Dynamic': 0.606827, 'Instruction Fetch Unit/Instruction Buffer/Runtime Dynamic': 0.0160044, 'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage': 0.00151885, 'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage with power gating': 0.000701682, 'Instruction Fetch Unit/Instruction Cache/Area': 3.14635, 'Instruction Fetch Unit/Instruction Cache/Gate Leakage': 0.029931, 'Instruction Fetch Unit/Instruction Cache/Peak Dynamic': 1.01802, 'Instruction Fetch Unit/Instruction Cache/Runtime Dynamic': 0.0388374, 'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage': 0.367022, 'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage with power gating': 0.180386, 'Instruction Fetch Unit/Instruction Decoder/Area': 1.85799, 'Instruction Fetch Unit/Instruction Decoder/Gate Leakage': 0.0222493, 'Instruction Fetch Unit/Instruction Decoder/Peak Dynamic': 1.37404, 'Instruction Fetch Unit/Instruction Decoder/Runtime Dynamic': 0.0543582, 'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage': 0.442943, 'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage with power gating': 0.166104, 'Instruction Fetch Unit/Peak Dynamic': 3.28594, 'Instruction Fetch Unit/Runtime Dynamic': 0.109982, 'Instruction Fetch Unit/Subthreshold Leakage': 0.932286, 'Instruction Fetch Unit/Subthreshold Leakage with power gating': 0.40843, 'L2/Area': 4.53318, 'L2/Gate Leakage': 0.015464, 'L2/Peak Dynamic': 0.0427043, 'L2/Runtime Dynamic': 0.0124827, 'L2/Subthreshold Leakage': 0.834142, 'L2/Subthreshold Leakage with power gating': 0.401066, 'Load Store Unit/Area': 8.80901, 'Load Store Unit/Data Cache/Area': 6.84535, 'Load Store Unit/Data Cache/Gate Leakage': 0.0279261, 'Load Store Unit/Data Cache/Peak Dynamic': 1.91467, 'Load Store Unit/Data Cache/Runtime Dynamic': 0.342768, 'Load Store Unit/Data Cache/Subthreshold Leakage': 0.527675, 'Load Store Unit/Data Cache/Subthreshold Leakage with power gating': 0.25085, 'Load Store Unit/Gate Leakage': 0.0350888, 'Load Store Unit/LoadQ/Area': 0.0836782, 'Load Store Unit/LoadQ/Gate Leakage': 0.00059896, 'Load Store Unit/LoadQ/Peak Dynamic': 0.0219204, 'Load Store Unit/LoadQ/Runtime Dynamic': 0.0219203, 'Load Store Unit/LoadQ/Subthreshold Leakage': 0.00941961, 'Load Store Unit/LoadQ/Subthreshold Leakage with power gating': 0.00536918, 'Load Store Unit/Peak Dynamic': 2.01818, 'Load Store Unit/Runtime Dynamic': 0.472792, 'Load Store Unit/StoreQ/Area': 0.322079, 'Load Store Unit/StoreQ/Gate Leakage': 0.00329971, 'Load Store Unit/StoreQ/Peak Dynamic': 0.054052, 'Load Store Unit/StoreQ/Runtime Dynamic': 0.108104, 'Load Store Unit/StoreQ/Subthreshold Leakage': 0.0345621, 'Load Store Unit/StoreQ/Subthreshold Leakage with power gating': 0.0197004, 'Load Store Unit/Subthreshold Leakage': 0.591321, 'Load Store Unit/Subthreshold Leakage with power gating': 0.283293, 'Memory Management Unit/Area': 0.4339, 'Memory Management Unit/Dtlb/Area': 0.0879726, 'Memory Management Unit/Dtlb/Gate Leakage': 0.00088729, 'Memory Management Unit/Dtlb/Peak Dynamic': 0.0191832, 'Memory Management Unit/Dtlb/Runtime Dynamic': 0.0198221, 'Memory Management Unit/Dtlb/Subthreshold Leakage': 0.0155699, 'Memory Management Unit/Dtlb/Subthreshold Leakage with power gating': 0.00887485, 'Memory Management Unit/Gate Leakage': 0.00808595, 'Memory Management Unit/Itlb/Area': 0.301552, 'Memory Management Unit/Itlb/Gate Leakage': 0.00393464, 'Memory Management Unit/Itlb/Peak Dynamic': 0.0632968, 'Memory Management Unit/Itlb/Runtime Dynamic': 0.00637403, 'Memory Management Unit/Itlb/Subthreshold Leakage': 0.0413758, 'Memory Management Unit/Itlb/Subthreshold Leakage with power gating': 0.0235842, 'Memory Management Unit/Peak Dynamic': 0.252359, 'Memory Management Unit/Runtime Dynamic': 0.0261961, 'Memory Management Unit/Subthreshold Leakage': 0.0766103, 'Memory Management Unit/Subthreshold Leakage with power gating': 0.0398333, 'Peak Dynamic': 13.3204, 'Renaming Unit/Area': 0.303608, 'Renaming Unit/FP Front End RAT/Area': 0.131045, 'Renaming Unit/FP Front End RAT/Gate Leakage': 0.00351123, 'Renaming Unit/FP Front End RAT/Peak Dynamic': 2.51468, 'Renaming Unit/FP Front End RAT/Runtime Dynamic': 0.0654206, 'Renaming Unit/FP Front End RAT/Subthreshold Leakage': 0.0308571, 'Renaming Unit/FP Front End RAT/Subthreshold Leakage with power gating': 0.0175885, 'Renaming Unit/Free List/Area': 0.0340654, 'Renaming Unit/Free List/Gate Leakage': 2.5481e-05, 'Renaming Unit/Free List/Peak Dynamic': 0.0306032, 'Renaming Unit/Free List/Runtime Dynamic': 0.00321753, 'Renaming Unit/Free List/Subthreshold Leakage': 0.000370144, 'Renaming Unit/Free List/Subthreshold Leakage with power gating': 0.000201064, 'Renaming Unit/Gate Leakage': 0.00708398, 'Renaming Unit/Int Front End RAT/Area': 0.0941223, 'Renaming Unit/Int Front End RAT/Gate Leakage': 0.000283242, 'Renaming Unit/Int Front End RAT/Peak Dynamic': 0.731965, 'Renaming Unit/Int Front End RAT/Runtime Dynamic': 0.0264284, 'Renaming Unit/Int Front End RAT/Subthreshold Leakage': 0.00435488, 'Renaming Unit/Int Front End RAT/Subthreshold Leakage with power gating': 0.00248228, 'Renaming Unit/Peak Dynamic': 3.58947, 'Renaming Unit/Runtime Dynamic': 0.0950665, 'Renaming Unit/Subthreshold Leakage': 0.0552466, 'Renaming Unit/Subthreshold Leakage with power gating': 0.0276461, 'Runtime Dynamic': 1.70472, 'Subthreshold Leakage': 6.16288, 'Subthreshold Leakage with power gating': 2.55328}, {'Area': 32.0201, 'Execution Unit/Area': 7.68434, 'Execution Unit/Complex ALUs/Area': 0.235435, 'Execution Unit/Complex ALUs/Gate Leakage': 0.0132646, 'Execution Unit/Complex ALUs/Peak Dynamic': 0.0232413, 'Execution Unit/Complex ALUs/Runtime Dynamic': 0.220943, 'Execution Unit/Complex ALUs/Subthreshold Leakage': 0.20111, 'Execution Unit/Complex ALUs/Subthreshold Leakage with power gating': 0.0754163, 'Execution Unit/Floating Point Units/Area': 4.6585, 'Execution Unit/Floating Point Units/Gate Leakage': 0.0656156, 'Execution Unit/Floating Point Units/Peak Dynamic': 0.13164, 'Execution Unit/Floating Point Units/Runtime Dynamic': 0.304033, 'Execution Unit/Floating Point Units/Subthreshold Leakage': 0.994829, 'Execution Unit/Floating Point Units/Subthreshold Leakage with power gating': 0.373061, 'Execution Unit/Gate Leakage': 0.120359, 'Execution Unit/Instruction Scheduler/Area': 1.66526, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Area': 0.275653, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Gate Leakage': 0.000977433, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Peak Dynamic': 1.04181, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Runtime Dynamic': 0.0536692, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage': 0.0143453, 'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage with power gating': 0.00810519, 'Execution Unit/Instruction Scheduler/Gate Leakage': 0.00568913, 'Execution Unit/Instruction Scheduler/Instruction Window/Area': 0.805223, 'Execution Unit/Instruction Scheduler/Instruction Window/Gate Leakage': 0.00414562, 'Execution Unit/Instruction Scheduler/Instruction Window/Peak Dynamic': 1.6763, 'Execution Unit/Instruction Scheduler/Instruction Window/Runtime Dynamic': 0.0865663, 'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage': 0.0625755, 'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage with power gating': 0.0355964, 'Execution Unit/Instruction Scheduler/Peak Dynamic': 3.82262, 'Execution Unit/Instruction Scheduler/ROB/Area': 0.584388, 'Execution Unit/Instruction Scheduler/ROB/Gate Leakage': 0.00056608, 'Execution Unit/Instruction Scheduler/ROB/Peak Dynamic': 1.10451, 'Execution Unit/Instruction Scheduler/ROB/Runtime Dynamic': 0.0436958, 'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage': 0.00906853, 'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage with power gating': 0.00364446, 'Execution Unit/Instruction Scheduler/Runtime Dynamic': 0.183931, 'Execution Unit/Instruction Scheduler/Subthreshold Leakage': 0.0859892, 'Execution Unit/Instruction Scheduler/Subthreshold Leakage with power gating': 0.047346, 'Execution Unit/Integer ALUs/Area': 0.47087, 'Execution Unit/Integer ALUs/Gate Leakage': 0.0265291, 'Execution Unit/Integer ALUs/Peak Dynamic': 0.0411987, 'Execution Unit/Integer ALUs/Runtime Dynamic': 0.101344, 'Execution Unit/Integer ALUs/Subthreshold Leakage': 0.40222, 'Execution Unit/Integer ALUs/Subthreshold Leakage with power gating': 0.150833, 'Execution Unit/Peak Dynamic': 4.13174, 'Execution Unit/Register Files/Area': 0.570804, 'Execution Unit/Register Files/Floating Point RF/Area': 0.208131, 'Execution Unit/Register Files/Floating Point RF/Gate Leakage': 0.000232788, 'Execution Unit/Register Files/Floating Point RF/Peak Dynamic': 0.0248697, 'Execution Unit/Register Files/Floating Point RF/Runtime Dynamic': 0.00225113, 'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage': 0.00399698, 'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage with power gating': 0.00176968, 'Execution Unit/Register Files/Gate Leakage': 0.000622708, 'Execution Unit/Register Files/Integer RF/Area': 0.362673, 'Execution Unit/Register Files/Integer RF/Gate Leakage': 0.00038992, 'Execution Unit/Register Files/Integer RF/Peak Dynamic': 0.0247301, 'Execution Unit/Register Files/Integer RF/Runtime Dynamic': 0.0166484, 'Execution Unit/Register Files/Integer RF/Subthreshold Leakage': 0.00614175, 'Execution Unit/Register Files/Integer RF/Subthreshold Leakage with power gating': 0.00246675, 'Execution Unit/Register Files/Peak Dynamic': 0.0495998, 'Execution Unit/Register Files/Runtime Dynamic': 0.0188996, 'Execution Unit/Register Files/Subthreshold Leakage': 0.0101387, 'Execution Unit/Register Files/Subthreshold Leakage with power gating': 0.00423643, 'Execution Unit/Results Broadcast Bus/Area Overhead': 0.0390912, 'Execution Unit/Results Broadcast Bus/Gate Leakage': 0.00537402, 'Execution Unit/Results Broadcast Bus/Peak Dynamic': 0.057911, 'Execution Unit/Results Broadcast Bus/Runtime Dynamic': 0.159052, 'Execution Unit/Results Broadcast Bus/Subthreshold Leakage': 0.081478, 'Execution Unit/Results Broadcast Bus/Subthreshold Leakage with power gating': 0.0305543, 'Execution Unit/Runtime Dynamic': 0.988203, 'Execution Unit/Subthreshold Leakage': 1.79543, 'Execution Unit/Subthreshold Leakage with power gating': 0.688821, 'Gate Leakage': 0.368936, 'Instruction Fetch Unit/Area': 5.85939, 'Instruction Fetch Unit/Branch Predictor/Area': 0.138516, 'Instruction Fetch Unit/Branch Predictor/Chooser/Area': 0.0435221, 'Instruction Fetch Unit/Branch Predictor/Chooser/Gate Leakage': 0.000278362, 'Instruction Fetch Unit/Branch Predictor/Chooser/Peak Dynamic': 0.0168831, 'Instruction Fetch Unit/Branch Predictor/Chooser/Runtime Dynamic': 4.39438e-05, 'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage': 0.00759719, 'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage with power gating': 0.0039236, 'Instruction Fetch Unit/Branch Predictor/Gate Leakage': 0.000757657, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Area': 0.0435221, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Gate Leakage': 0.000278362, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Peak Dynamic': 0.0168831, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Runtime Dynamic': 4.39438e-05, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage': 0.00759719, 'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage with power gating': 0.0039236, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Area': 0.0257064, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Gate Leakage': 0.000154548, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Peak Dynamic': 0.0142575, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Runtime Dynamic': 3.8421e-05, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage': 0.00384344, 'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage with power gating': 0.00198631, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Area': 0.0151917, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Gate Leakage': 8.00196e-05, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Peak Dynamic': 0.00527447, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Runtime Dynamic': 1.49532e-05, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage': 0.00181347, 'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage with power gating': 0.000957045, 'Instruction Fetch Unit/Branch Predictor/Peak Dynamic': 0.0597838, 'Instruction Fetch Unit/Branch Predictor/RAS/Area': 0.0105732, 'Instruction Fetch Unit/Branch Predictor/RAS/Gate Leakage': 4.63858e-05, 'Instruction Fetch Unit/Branch Predictor/RAS/Peak Dynamic': 0.0117602, 'Instruction Fetch Unit/Branch Predictor/RAS/Runtime Dynamic': 0.000239156, 'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage': 0.000932505, 'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage with power gating': 0.000494733, 'Instruction Fetch Unit/Branch Predictor/Runtime Dynamic': 0.000365465, 'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage': 0.0199703, 'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage with power gating': 0.0103282, 'Instruction Fetch Unit/Branch Target Buffer/Area': 0.64954, 'Instruction Fetch Unit/Branch Target Buffer/Gate Leakage': 0.00272758, 'Instruction Fetch Unit/Branch Target Buffer/Peak Dynamic': 0.177867, 'Instruction Fetch Unit/Branch Target Buffer/Runtime Dynamic': 0.000416114, 'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage': 0.0811682, 'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage with power gating': 0.0435357, 'Instruction Fetch Unit/Gate Leakage': 0.0589979, 'Instruction Fetch Unit/Instruction Buffer/Area': 0.0226323, 'Instruction Fetch Unit/Instruction Buffer/Gate Leakage': 6.83558e-05, 'Instruction Fetch Unit/Instruction Buffer/Peak Dynamic': 0.606827, 'Instruction Fetch Unit/Instruction Buffer/Runtime Dynamic': 0.0160046, 'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage': 0.00151885, 'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage with power gating': 0.000701682, 'Instruction Fetch Unit/Instruction Cache/Area': 3.14635, 'Instruction Fetch Unit/Instruction Cache/Gate Leakage': 0.029931, 'Instruction Fetch Unit/Instruction Cache/Peak Dynamic': 1.01803, 'Instruction Fetch Unit/Instruction Cache/Runtime Dynamic': 0.0388377, 'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage': 0.367022, 'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage with power gating': 0.180386, 'Instruction Fetch Unit/Instruction Decoder/Area': 1.85799, 'Instruction Fetch Unit/Instruction Decoder/Gate Leakage': 0.0222493, 'Instruction Fetch Unit/Instruction Decoder/Peak Dynamic': 1.37404, 'Instruction Fetch Unit/Instruction Decoder/Runtime Dynamic': 0.0543588, 'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage': 0.442943, 'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage with power gating': 0.166104, 'Instruction Fetch Unit/Peak Dynamic': 3.28595, 'Instruction Fetch Unit/Runtime Dynamic': 0.109983, 'Instruction Fetch Unit/Subthreshold Leakage': 0.932286, 'Instruction Fetch Unit/Subthreshold Leakage with power gating': 0.40843, 'L2/Area': 4.53318, 'L2/Gate Leakage': 0.015464, 'L2/Peak Dynamic': 0.0427043, 'L2/Runtime Dynamic': 0.0124828, 'L2/Subthreshold Leakage': 0.834142, 'L2/Subthreshold Leakage with power gating': 0.401066, 'Load Store Unit/Area': 8.80901, 'Load Store Unit/Data Cache/Area': 6.84535, 'Load Store Unit/Data Cache/Gate Leakage': 0.0279261, 'Load Store Unit/Data Cache/Peak Dynamic': 1.91468, 'Load Store Unit/Data Cache/Runtime Dynamic': 0.342771, 'Load Store Unit/Data Cache/Subthreshold Leakage': 0.527675, 'Load Store Unit/Data Cache/Subthreshold Leakage with power gating': 0.25085, 'Load Store Unit/Gate Leakage': 0.0350888, 'Load Store Unit/LoadQ/Area': 0.0836782, 'Load Store Unit/LoadQ/Gate Leakage': 0.00059896, 'Load Store Unit/LoadQ/Peak Dynamic': 0.0219206, 'Load Store Unit/LoadQ/Runtime Dynamic': 0.0219205, 'Load Store Unit/LoadQ/Subthreshold Leakage': 0.00941961, 'Load Store Unit/LoadQ/Subthreshold Leakage with power gating': 0.00536918, 'Load Store Unit/Peak Dynamic': 2.01819, 'Load Store Unit/Runtime Dynamic': 0.472796, 'Load Store Unit/StoreQ/Area': 0.322079, 'Load Store Unit/StoreQ/Gate Leakage': 0.00329971, 'Load Store Unit/StoreQ/Peak Dynamic': 0.0540526, 'Load Store Unit/StoreQ/Runtime Dynamic': 0.108105, 'Load Store Unit/StoreQ/Subthreshold Leakage': 0.0345621, 'Load Store Unit/StoreQ/Subthreshold Leakage with power gating': 0.0197004, 'Load Store Unit/Subthreshold Leakage': 0.591321, 'Load Store Unit/Subthreshold Leakage with power gating': 0.283293, 'Memory Management Unit/Area': 0.4339, 'Memory Management Unit/Dtlb/Area': 0.0879726, 'Memory Management Unit/Dtlb/Gate Leakage': 0.00088729, 'Memory Management Unit/Dtlb/Peak Dynamic': 0.0191834, 'Memory Management Unit/Dtlb/Runtime Dynamic': 0.0198223, 'Memory Management Unit/Dtlb/Subthreshold Leakage': 0.0155699, 'Memory Management Unit/Dtlb/Subthreshold Leakage with power gating': 0.00887485, 'Memory Management Unit/Gate Leakage': 0.00808595, 'Memory Management Unit/Itlb/Area': 0.301552, 'Memory Management Unit/Itlb/Gate Leakage': 0.00393464, 'Memory Management Unit/Itlb/Peak Dynamic': 0.0632972, 'Memory Management Unit/Itlb/Runtime Dynamic': 0.00637408, 'Memory Management Unit/Itlb/Subthreshold Leakage': 0.0413758, 'Memory Management Unit/Itlb/Subthreshold Leakage with power gating': 0.0235842, 'Memory Management Unit/Peak Dynamic': 0.25236, 'Memory Management Unit/Runtime Dynamic': 0.0261964, 'Memory Management Unit/Subthreshold Leakage': 0.0766103, 'Memory Management Unit/Subthreshold Leakage with power gating': 0.0398333, 'Peak Dynamic': 13.3204, 'Renaming Unit/Area': 0.303608, 'Renaming Unit/FP Front End RAT/Area': 0.131045, 'Renaming Unit/FP Front End RAT/Gate Leakage': 0.00351123, 'Renaming Unit/FP Front End RAT/Peak Dynamic': 2.51468, 'Renaming Unit/FP Front End RAT/Runtime Dynamic': 0.0654209, 'Renaming Unit/FP Front End RAT/Subthreshold Leakage': 0.0308571, 'Renaming Unit/FP Front End RAT/Subthreshold Leakage with power gating': 0.0175885, 'Renaming Unit/Free List/Area': 0.0340654, 'Renaming Unit/Free List/Gate Leakage': 2.5481e-05, 'Renaming Unit/Free List/Peak Dynamic': 0.0306032, 'Renaming Unit/Free List/Runtime Dynamic': 0.00321756, 'Renaming Unit/Free List/Subthreshold Leakage': 0.000370144, 'Renaming Unit/Free List/Subthreshold Leakage with power gating': 0.000201064, 'Renaming Unit/Gate Leakage': 0.00708398, 'Renaming Unit/Int Front End RAT/Area': 0.0941223, 'Renaming Unit/Int Front End RAT/Gate Leakage': 0.000283242, 'Renaming Unit/Int Front End RAT/Peak Dynamic': 0.731965, 'Renaming Unit/Int Front End RAT/Runtime Dynamic': 0.0264287, 'Renaming Unit/Int Front End RAT/Subthreshold Leakage': 0.00435488, 'Renaming Unit/Int Front End RAT/Subthreshold Leakage with power gating': 0.00248228, 'Renaming Unit/Peak Dynamic': 3.58947, 'Renaming Unit/Runtime Dynamic': 0.0950671, 'Renaming Unit/Subthreshold Leakage': 0.0552466, 'Renaming Unit/Subthreshold Leakage with power gating': 0.0276461, 'Runtime Dynamic': 1.70473, 'Subthreshold Leakage': 6.16288, 'Subthreshold Leakage with power gating': 2.55328}], 'DRAM': {'Area': 0, 'Gate Leakage': 0, 'Peak Dynamic': 8.59419502088197, 'Runtime Dynamic': 8.59419502088197, 'Subthreshold Leakage': 4.252, 'Subthreshold Leakage with power gating': 4.252}, 'L3': [{'Area': 61.9075, 'Gate Leakage': 0.0484137, 'Peak Dynamic': 0.45468, 'Runtime Dynamic': 0.179204, 'Subthreshold Leakage': 6.80085, 'Subthreshold Leakage with power gating': 3.32364}], 'Processor': {'Area': 191.908, 'Gate Leakage': 1.53485, 'Peak Dynamic': 60.4483, 'Peak Power': 93.5606, 'Runtime Dynamic': 9.52364, 'Subthreshold Leakage': 31.5774, 'Subthreshold Leakage with power gating': 13.9484, 'Total Cores/Area': 128.669, 'Total Cores/Gate Leakage': 1.4798, 'Total Cores/Peak Dynamic': 59.9937, 'Total Cores/Runtime Dynamic': 9.34443, 'Total Cores/Subthreshold Leakage': 24.7074, 'Total Cores/Subthreshold Leakage with power gating': 10.2429, 'Total L3s/Area': 61.9075, 'Total L3s/Gate Leakage': 0.0484137, 'Total L3s/Peak Dynamic': 0.45468, 'Total L3s/Runtime Dynamic': 0.179204, 'Total L3s/Subthreshold Leakage': 6.80085, 'Total L3s/Subthreshold Leakage with power gating': 3.32364, 'Total Leakage': 33.1122, 'Total NoCs/Area': 1.33155, 'Total NoCs/Gate Leakage': 0.00662954, 'Total NoCs/Peak Dynamic': 0.0, 'Total NoCs/Runtime Dynamic': 0.0, 'Total NoCs/Subthreshold Leakage': 0.0691322, 'Total NoCs/Subthreshold Leakage with power gating': 0.0259246}}
75.126915
124
0.682157
8,095
68,666
5.780482
0.065967
0.123437
0.112837
0.093347
0.942406
0.934541
0.922168
0.894065
0.871861
0.854401
0
0.132389
0.224143
68,666
914
125
75.126915
0.745941
0
0
0.66849
0
0
0.656866
0.048058
0
0
0
0
0
1
0
false
0
0
0
0
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
d545220585c8ef43745228e1c977b6fe55a30efe
2,829
py
Python
tests/unit/test_source/test_tsv_source.py
STARInformatics/kgx
a4b4dd73f866486466c03579f62e0c527ef1af1e
[ "BSD-3-Clause" ]
null
null
null
tests/unit/test_source/test_tsv_source.py
STARInformatics/kgx
a4b4dd73f866486466c03579f62e0c527ef1af1e
[ "BSD-3-Clause" ]
null
null
null
tests/unit/test_source/test_tsv_source.py
STARInformatics/kgx
a4b4dd73f866486466c03579f62e0c527ef1af1e
[ "BSD-3-Clause" ]
null
null
null
import os import pprint import pytest from kgx.source import TsvSource from tests import RESOURCE_DIR def test_read_tsv(): """ Read a TSV using TsvSource. """ s = TsvSource() g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test_nodes.tsv'), format='tsv') nodes = [] for rec in g: if rec: nodes.append(rec) assert len(nodes) == 3 nodes.sort() n1 = nodes.pop()[-1] assert n1['id'] == 'CURIE:456' assert n1['name'] == 'Disease 456' assert 'biolink:Disease' in n1['category'] assert n1['description'] == '"Node of type Disease, CURIE:456"' g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test_edges.tsv'), format='tsv') edges = [] for rec in g: if rec: edges.append(rec) e1 = edges.pop()[-1] assert 'id' in e1 assert e1['subject'] == 'CURIE:123' assert e1['object'] == 'CURIE:456' assert e1['predicate'] == 'biolink:related_to' assert e1['relation'] == 'biolink:related_to' assert 'PMID:1' in e1['publications'] def test_read_csv(): """ Read a CSV using TsvSource. """ s = TsvSource() g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test_nodes.csv'), format='csv') nodes = [] for rec in g: if rec: nodes.append(rec) assert len(nodes) == 3 nodes.sort() n1 = nodes.pop()[-1] assert n1['id'] == 'CURIE:456' assert n1['name'] == 'Disease 456' assert 'biolink:Disease' in n1['category'] assert n1['description'] == 'Node of type Disease, CURIE:456' g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test_edges.csv'), format='csv') edges = [] for rec in g: if rec: print(rec) edges.append(rec) e1 = edges.pop()[-1] assert 'id' in e1 assert e1['subject'] == 'CURIE:123' assert e1['object'] == 'CURIE:456' assert e1['predicate'] == 'biolink:related_to' assert e1['relation'] == 'biolink:related_to' assert 'PMID:1' in e1['publications'] def test_read_tsv_compressed(): """ Read a compressed TSV archive using TsvSource. """ s = TsvSource() g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test.tar'), format='tsv', compression='tar') nodes = [] edges = [] for rec in g: if rec: if len(rec) == 4: edges.append(rec) else: nodes.append(nodes) assert len(nodes) == 3 assert len(edges) == 1 g = s.parse( filename=os.path.join(RESOURCE_DIR, 'test.tar.gz'), format='tsv', compression='tar.gz' ) nodes = [] edges = [] for rec in g: if rec: if len(rec) == 4: edges.append(rec) else: nodes.append(nodes) assert len(nodes) == 3 assert len(edges) == 1
26.942857
97
0.563097
378
2,829
4.156085
0.179894
0.040738
0.026735
0.057288
0.8345
0.8345
0.8345
0.812222
0.812222
0.812222
0
0.033382
0.279958
2,829
104
98
27.201923
0.73785
0.036055
0
0.77381
0
0
0.18426
0
0
0
0
0
0.309524
1
0.035714
false
0
0.059524
0
0.095238
0.02381
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
638c3d20dddafd292eba1978584a5a3502ee86ef
24,461
py
Python
tmapi/tests/indices/test_scoped_index.py
ajenhl/django-tmapi
02f009e1b508218cf330ca7748c3a1dd110f3e8d
[ "Apache-2.0" ]
2
2015-03-22T03:23:36.000Z
2017-01-08T10:57:18.000Z
tmapi/tests/indices/test_scoped_index.py
ajenhl/django-tmapi
02f009e1b508218cf330ca7748c3a1dd110f3e8d
[ "Apache-2.0" ]
null
null
null
tmapi/tests/indices/test_scoped_index.py
ajenhl/django-tmapi
02f009e1b508218cf330ca7748c3a1dd110f3e8d
[ "Apache-2.0" ]
1
2020-12-28T04:40:34.000Z
2020-12-28T04:40:34.000Z
# Copyright 2011 Jamie Norrish (jamie@artefact.org.nz) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Module containing tests against the `ScopedIndex` interface. Most if not all of these tests are ported from the public domain tests that come with the TMAPI 2.0 distribution (http://www.tmapi.org/2.0/). """ from tmapi.exceptions import IllegalArgumentException from tmapi.indices.scoped_index import ScopedIndex from tmapi.tests.models.tmapi_test_case import TMAPITestCase class ScopedIndexTest (TMAPITestCase): def setUp (self): super(ScopedIndexTest, self).setUp() self._index = self.tm.get_index(ScopedIndex) self._index.open() def tearDown (self): super(ScopedIndexTest, self).tearDown() self._index.close() def _update_index (self): if not self._index.is_auto_updated(): self._index.reindex() def test_association (self): theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_associations().count()) self.assertEqual(0, self._index.get_associations(theme).count()) self.assertEqual(0, self._index.get_association_themes().count()) scoped = self.create_association() self.assertEqual(0, scoped.get_scope().count()) self._update_index() self.assertEqual(1, self._index.get_associations().count()) self.assertTrue(scoped in self._index.get_associations()) self.assertFalse(theme in self._index.get_association_themes()) scoped.add_theme(theme) self._update_index() self.assertEqual(0, self._index.get_associations().count()) self.assertFalse(scoped in self._index.get_associations()) self.assertNotEqual(0, self._index.get_association_themes().count()) self.assertEqual(1, self._index.get_association_themes().count()) self.assertTrue(scoped in self._index.get_associations(theme)) self.assertTrue(theme in self._index.get_association_themes()) scoped.remove() self._update_index() self.assertEqual(0, self._index.get_associations().count()) self.assertFalse(scoped in self._index.get_associations()) self.assertFalse(theme in self._index.get_association_themes()) def test_association_match_all (self): theme = self.create_topic() theme2 = self.create_topic() unused_theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_associations().count()) self.assertEqual(0, self._index.get_associations(theme).count()) self.assertEqual(0, self._index.get_association_themes().count()) scoped = self.create_association() self.assertEqual(0, scoped.get_scope().count()) self._update_index() self.assertEqual(1, self._index.get_associations().count()) self.assertTrue(scoped in self._index.get_associations()) self.assertFalse(theme in self._index.get_association_themes()) scoped.add_theme(theme) self._update_index() self.assertEqual(1, self._index.get_association_themes().count()) self.assertTrue(scoped in self._index.get_associations([theme], True)) self.assertTrue(scoped in self._index.get_associations([theme], False)) scoped.add_theme(theme2) self._update_index() self.assertEqual(2, self._index.get_association_themes().count()) self.assertTrue(scoped in self._index.get_associations([theme], True)) self.assertTrue(scoped in self._index.get_associations([theme], False)) self.assertTrue(scoped in self._index.get_associations([theme2], True)) self.assertTrue(scoped in self._index.get_associations([theme2], False)) self.assertTrue(scoped in self._index.get_associations( [theme, theme2], False)) self.assertTrue(scoped in self._index.get_associations( [theme, theme2], True)) self.assertTrue(scoped in self._index.get_associations( [theme, unused_theme], False)) self.assertTrue(scoped in self._index.get_associations( [theme2, unused_theme], False)) self.assertFalse(scoped in self._index.get_associations( [theme, unused_theme], True)) self.assertFalse(scoped in self._index.get_associations( [theme2, unused_theme], True)) def test_association_match_all_illegal (self): self.assertRaises(IllegalArgumentException, self._index.get_associations, None, True) def test_occurrence (self): theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_occurrences().count()) self.assertEqual(0, self._index.get_occurrences(theme).count()) self.assertEqual(0, self._index.get_occurrence_themes().count()) scoped = self.create_occurrence() self.assertEqual(0, scoped.get_scope().count()) self._update_index() self.assertEqual(1, self._index.get_occurrences().count()) self.assertTrue(scoped in self._index.get_occurrences()) self.assertFalse(theme in self._index.get_occurrence_themes()) scoped.add_theme(theme) self._update_index() self.assertEqual(0, self._index.get_occurrences().count()) self.assertFalse(scoped in self._index.get_occurrences()) self.assertNotEqual(0, self._index.get_occurrence_themes().count()) self.assertEqual(1, self._index.get_occurrence_themes().count()) self.assertTrue(scoped in self._index.get_occurrences(theme)) self.assertTrue(theme in self._index.get_occurrence_themes()) scoped.remove() self._update_index() self.assertEqual(0, self._index.get_occurrences().count()) self.assertFalse(scoped in self._index.get_occurrences()) self.assertFalse(theme in self._index.get_occurrence_themes()) def test_occurrence_match_all (self): theme = self.create_topic() theme2 = self.create_topic() unused_theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_occurrences().count()) self.assertEqual(0, self._index.get_occurrences(theme).count()) self.assertEqual(0, self._index.get_occurrence_themes().count()) scoped = self.create_occurrence() self.assertEqual(0, scoped.get_scope().count()) self._update_index() self.assertEqual(1, self._index.get_occurrences().count()) self.assertTrue(scoped in self._index.get_occurrences()) self.assertFalse(theme in self._index.get_occurrence_themes()) scoped.add_theme(theme) self._update_index() self.assertEqual(1, self._index.get_occurrence_themes().count()) self.assertTrue(scoped in self._index.get_occurrences([theme], True)) self.assertTrue(scoped in self._index.get_occurrences([theme], False)) scoped.add_theme(theme2) self._update_index() self.assertEqual(2, self._index.get_occurrence_themes().count()) self.assertTrue(scoped in self._index.get_occurrences([theme], True)) self.assertTrue(scoped in self._index.get_occurrences([theme], False)) self.assertTrue(scoped in self._index.get_occurrences([theme2], True)) self.assertTrue(scoped in self._index.get_occurrences([theme2], False)) self.assertTrue(scoped in self._index.get_occurrences( [theme, theme2], False)) self.assertTrue(scoped in self._index.get_occurrences( [theme, theme2], True)) self.assertTrue(scoped in self._index.get_occurrences( [theme, unused_theme], False)) self.assertTrue(scoped in self._index.get_occurrences( [theme2, unused_theme], False)) self.assertFalse(scoped in self._index.get_occurrences( [theme, unused_theme], True)) self.assertFalse(scoped in self._index.get_occurrences( [theme2, unused_theme], True)) def test_occurrence_match_all_illegal (self): self.assertRaises(IllegalArgumentException, self._index.get_occurrences, None, True) def test_name (self): theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_names().count()) self.assertEqual(0, self._index.get_names(theme).count()) self.assertEqual(0, self._index.get_name_themes().count()) scoped = self.create_name() self.assertEqual(0, scoped.get_scope().count()) self._update_index() self.assertEqual(1, self._index.get_names().count()) self.assertTrue(scoped in self._index.get_names()) self.assertFalse(theme in self._index.get_name_themes()) scoped.add_theme(theme) self._update_index() self.assertEqual(0, self._index.get_names().count()) self.assertFalse(scoped in self._index.get_names()) self.assertNotEqual(0, self._index.get_name_themes().count()) self.assertEqual(1, self._index.get_name_themes().count()) self.assertTrue(scoped in self._index.get_names(theme)) self.assertTrue(theme in self._index.get_name_themes()) scoped.remove() self._update_index() self.assertEqual(0, self._index.get_names().count()) self.assertFalse(scoped in self._index.get_names()) self.assertFalse(theme in self._index.get_name_themes()) def test_name_2 (self): theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_names().count()) self.assertEqual(0, self._index.get_names(theme).count()) self.assertEqual(0, self._index.get_name_themes().count()) scoped = self.create_topic().create_name('tinyTiM', scope=theme) self.assertEqual(1, scoped.get_scope().count()) self._update_index() self.assertEqual(0, self._index.get_names().count()) self.assertFalse(scoped in self._index.get_names()) self.assertNotEqual(0, self._index.get_name_themes().count()) self.assertEqual(1, self._index.get_name_themes().count()) self.assertTrue(scoped in self._index.get_names(theme)) self.assertTrue(theme in self._index.get_name_themes()) scoped.remove() self._update_index() self.assertEqual(0, self._index.get_names().count()) self.assertFalse(scoped in self._index.get_names()) self.assertEqual(0, self._index.get_names(theme).count()) self.assertFalse(theme in self._index.get_name_themes()) def test_name_match_all (self): theme = self.create_topic() theme2 = self.create_topic() unused_theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_names().count()) self.assertEqual(0, self._index.get_names(theme).count()) self.assertEqual(0, self._index.get_name_themes().count()) scoped = self.create_name() self.assertEqual(0, scoped.get_scope().count()) self._update_index() self.assertEqual(1, self._index.get_names().count()) self.assertTrue(scoped in self._index.get_names()) self.assertFalse(theme in self._index.get_name_themes()) scoped.add_theme(theme) self._update_index() self.assertEqual(1, self._index.get_name_themes().count()) self.assertTrue(scoped in self._index.get_names([theme], True)) self.assertTrue(scoped in self._index.get_names([theme], False)) scoped.add_theme(theme2) self._update_index() self.assertEqual(2, self._index.get_name_themes().count()) self.assertTrue(scoped in self._index.get_names([theme], True)) self.assertTrue(scoped in self._index.get_names([theme], False)) self.assertTrue(scoped in self._index.get_names([theme2], True)) self.assertTrue(scoped in self._index.get_names([theme2], False)) self.assertTrue(scoped in self._index.get_names([theme, theme2], True)) self.assertTrue(scoped in self._index.get_names([theme, theme2], False)) self.assertTrue(scoped in self._index.get_names( [theme, unused_theme], False)) self.assertTrue(scoped in self._index.get_names( [theme2, unused_theme], False)) self.assertFalse(scoped in self._index.get_names( [theme, unused_theme], True)) self.assertFalse(scoped in self._index.get_names( [theme2, unused_theme], True)) def test_name_match_all_illegal (self): self.assertRaises(IllegalArgumentException, self._index.get_names, None, True) def test_variant_illegal (self): self.assertRaises(IllegalArgumentException, self._index.get_variants, None) def test_variant_match_all_illegal (self): self.assertRaises(IllegalArgumentException, self._index.get_variants, None, True) def test_variant (self): theme = self.create_topic() theme2 = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_variants(theme).count()) self.assertEqual(0, self._index.get_variant_themes().count()) name = self.create_name() self.assertEqual(0, name.get_scope().count()) scoped = name.create_variant('Variant', theme) self.assertEqual(1, scoped.get_scope().count(), 'Unexpected variant\'s scope size') self._update_index() self.assertNotEqual(0, self._index.get_variant_themes().count()) self.assertEqual(1, self._index.get_variant_themes().count(), 'Unexpected number of variant themes') self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) name.add_theme(theme2) self.assertEqual(1, name.get_scope().count()) self.assertEqual(2, scoped.get_scope().count(), 'The scope change of the parent is not reflected in the variant\'s scope') self._update_index() self.assertEqual(2, self._index.get_variant_themes().count(), 'Change of the parent\'s scope is not reflected in the index') self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) self.assertTrue(scoped in self._index.get_variants(theme2)) self.assertTrue(theme2 in self._index.get_variant_themes()) name.remove_theme(theme2) self._update_index() self.assertNotEqual(0, self._index.get_variant_themes().count()) self.assertEqual(1, self._index.get_variant_themes().count(), 'The scope change in the name is not reflected in the variant') self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) scoped.add_theme(theme2) self._update_index() self.assertEqual(2, self._index.get_variant_themes().count(), 'Change of the variant\'s scope is not reflected in the index') self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) self.assertTrue(scoped in self._index.get_variants(theme2)) self.assertTrue(theme2 in self._index.get_variant_themes()) name.add_theme(theme2) self._update_index() self.assertEqual(2, self._index.get_variant_themes().count(), 'Adding a theme to the variant\'s parent is not reflected in the index') self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) self.assertTrue(scoped in self._index.get_variants(theme2)) self.assertTrue(theme2 in self._index.get_variant_themes()) name.remove_theme(theme2) self._update_index() self.assertEqual(2, self._index.get_variant_themes().count(), 'Removing the name\'s theme MUST NOT be reflected in the variant\'s scope') self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) self.assertTrue(scoped in self._index.get_variants(theme2)) self.assertTrue(theme2 in self._index.get_variant_themes()) scoped.remove_theme(theme2) self.assertNotEqual(0, self._index.get_variant_themes().count()) self.assertEqual(1, self._index.get_variant_themes().count()) self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) def test_variant_2 (self): theme = self.create_topic() theme2 = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_variants(theme).count()) self.assertEqual(0, self._index.get_variants(theme2).count()) self.assertEqual(0, self._index.get_variant_themes().count()) name = self.create_topic().create_name('Name', scope=theme2) self.assertEqual(1, name.get_scope().count()) scoped = name.create_variant('Variant', theme) self.assertEqual(2, scoped.get_scope().count()) self._update_index() self.assertEqual(2, self._index.get_variant_themes().count()) self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) self.assertTrue(scoped in self._index.get_variants(theme2)) self.assertTrue(theme2 in self._index.get_variant_themes()) name.remove_theme(theme2) self.assertEqual(0, name.get_scope().count()) self._update_index() self.assertEqual(1, self._index.get_variant_themes().count()) self.assertTrue(scoped in self._index.get_variants(theme)) self.assertTrue(theme in self._index.get_variant_themes()) def test_variant_match_all (self): theme = self.create_topic() theme2 = self.create_topic() unused_theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_variants(theme).count()) self.assertEqual(0, self._index.get_variants(theme2).count()) self.assertEqual(0, self._index.get_variant_themes().count()) name = self.create_topic().create_name('Name') self.assertEqual(0, name.get_scope().count()) scoped = name.create_variant('Variant', theme) self.assertEqual(1, scoped.get_scope().count()) self._update_index() self.assertEqual(1, self._index.get_variant_themes().count()) self.assertTrue(scoped in self._index.get_variants([theme], True)) self.assertTrue(scoped in self._index.get_variants([theme], False)) self.assertFalse(scoped in self._index.get_variants([theme2], True)) self.assertFalse(scoped in self._index.get_variants([theme2], False)) scoped.add_theme(theme2) self._update_index() self.assertTrue(scoped in self._index.get_variants([theme], True)) self.assertTrue(scoped in self._index.get_variants([theme], False)) self.assertTrue(scoped in self._index.get_variants([theme2], True)) self.assertTrue(scoped in self._index.get_variants([theme2], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2], True)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2, unused_theme], False)) self.assertFalse(scoped in self._index.get_variants( [theme, theme2, unused_theme], True)) name_theme = self.create_topic() name.add_theme(name_theme) self._update_index() self.assertTrue(scoped in self._index.get_variants([theme], True)) self.assertTrue(scoped in self._index.get_variants([theme], False)) self.assertTrue(scoped in self._index.get_variants([theme2], True)) self.assertTrue(scoped in self._index.get_variants([theme2], False)) self.assertTrue(scoped in self._index.get_variants([name_theme], True)) self.assertTrue(scoped in self._index.get_variants([name_theme], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2], True)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2, name_theme], True)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2, name_theme], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2, unused_theme], False)) self.assertFalse(scoped in self._index.get_variants( [theme, theme2, unused_theme], True)) name.remove_theme(name_theme) self._update_index() self.assertFalse(scoped in self._index.get_variants([name_theme], True)) self.assertFalse(scoped in self._index.get_variants( [name_theme], False)) self.assertFalse(scoped in self._index.get_variants( [theme, theme2, name_theme], True)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2, name_theme], False)) scoped.remove_theme(theme) self._update_index() self.assertFalse(scoped in self._index.get_variants([theme], True)) self.assertFalse(scoped in self._index.get_variants([theme], False)) self.assertFalse(scoped in self._index.get_variants( [theme, theme2], True)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2], False)) def test_variant_match_all_2 (self): theme = self.create_topic() theme2 = self.create_topic() unused_theme = self.create_topic() name_theme = self.create_topic() self._update_index() self.assertEqual(0, self._index.get_variants(theme).count()) self.assertEqual(0, self._index.get_variants(theme2).count()) self.assertEqual(0, self._index.get_variant_themes().count()) name = self.create_topic().create_name('Name', scope=name_theme) self.assertEqual(1, name.get_scope().count()) scoped = name.create_variant('Variant', [theme, theme2]) self.assertEqual(3, scoped.get_scope().count()) self._update_index() self.assertEqual(3, self._index.get_variant_themes().count()) self.assertTrue(scoped in self._index.get_variants([theme], True)) self.assertTrue(scoped in self._index.get_variants([theme], False)) self.assertTrue(scoped in self._index.get_variants([theme2], True)) self.assertTrue(scoped in self._index.get_variants([theme2], False)) self.assertTrue(scoped in self._index.get_variants([name_theme], True)) self.assertTrue(scoped in self._index.get_variants([name_theme], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2], True)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2, name_theme], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2, name_theme], False)) self.assertTrue(scoped in self._index.get_variants( [theme, theme2, unused_theme], False)) self.assertFalse(scoped in self._index.get_variants( [theme, theme2, unused_theme], True))
52.717672
145
0.673685
3,022
24,461
5.196889
0.049967
0.12894
0.168099
0.124801
0.915123
0.903725
0.896721
0.885132
0.875963
0.859472
0
0.009638
0.206778
24,461
463
146
52.831533
0.799773
0.031847
0
0.805226
0
0
0.013608
0
0
0
0
0
0.565321
1
0.045131
false
0
0.007126
0
0.054632
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
1
0
0
0
0
0
0
0
0
0
8
98460f90297c5bb37bad032f1201cc32789bbc1a
56,611
py
Python
adapters/adapter.py
kweveen/tng-vnv-platform-adapter
90b8499bb564664a751af26ff20f7ddbd5358b6a
[ "Apache-2.0" ]
null
null
null
adapters/adapter.py
kweveen/tng-vnv-platform-adapter
90b8499bb564664a751af26ff20f7ddbd5358b6a
[ "Apache-2.0" ]
null
null
null
adapters/adapter.py
kweveen/tng-vnv-platform-adapter
90b8499bb564664a751af26ff20f7ddbd5358b6a
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/python from flask import Flask, request, jsonify, render_template import os, sys, logging, uuid, json from werkzeug import secure_filename #import serviceplatform import psycopg2 import requests import subprocess import models.database as database import re import ast from ast import literal_eval FILE = "db-config.cfg" class Adapter: def __init__(self, name): self.name = name self.host = "host" self.type = "type" def getName(self): return self.name def setName(self, newName): self.name = newName def getHost(self): return self.host def setHost(self, newHost): self.host = newHost def getType(self): return self.type def setType(self, newType): self.type = newType def updateToken(self,token): try: db = database.Database(FILE) connection = psycopg2.connect(user = db.user, password = db.password, host = db.host, port = db.port, database = db.database) cursor = connection.cursor() print ( connection.get_dsn_parameters(),"\n") print (self.name) get_type = "SELECT type FROM service_platforms WHERE name=\'" +self.name+ "\'" print (get_type) update_token = "UPDATE service_platforms SET service_token = \'" +token+ "\' WHERE name = \'" +self.name+ "\'" print (update_token) cursor.execute(update_token) connection.commit() return "token updated", 200 except (Exception, psycopg2.Error) as error : print (error) exception_message = str(error) return exception_message, 401 finally: #closing database connection. if(connection): cursor.close() connection.close() print("PostgreSQL connection is closed") def getDBType(self): try: db = database.Database(FILE) connection = psycopg2.connect(user = db.user, password = db.password, host = db.host, port = db.port, database = db.database) cursor = connection.cursor() print ( connection.get_dsn_parameters(),"\n") #create table Service Platforms get_type = "SELECT type FROM service_platforms WHERE name=\'" +self.name+ "\'" print (get_type) cursor.execute(get_type) all = cursor.fetchall() #return jsonify(all), 200 type_0 = all.__str__() print(type_0) type_1 = type_0[3:] print(type_1) type_2 = type_1[:-4] print(type_2) return type_2 except (Exception, psycopg2.Error) as error : print (error) exception_message = str(error) return exception_message, 401 finally: #closing database connection. if(connection): cursor.close() connection.close() print("PostgreSQL connection is closed") def getDBUserName(self): try: db = database.Database(FILE) connection = psycopg2.connect(user = db.user, password = db.password, host = db.host, port = db.port, database = db.database) cursor = connection.cursor() print ( connection.get_dsn_parameters(),"\n") #create table Service Platforms get_username = "SELECT username FROM service_platforms WHERE name=\'" +self.name+ "\'" print (get_username) cursor.execute(get_username) all = cursor.fetchall() #return jsonify(all), 200 type_0 = all.__str__() print(type_0) type_1 = type_0[3:] print(type_1) type_2 = type_1[:-4] print(type_2) return type_2 except (Exception, psycopg2.Error) as error : print (error) exception_message = str(error) return exception_message, 401 finally: #closing database connection. if(connection): cursor.close() connection.close() print("PostgreSQL connection is closed") def getDBProjectName(self): try: db = database.Database(FILE) connection = psycopg2.connect(user = db.user, password = db.password, host = db.host, port = db.port, database = db.database) cursor = connection.cursor() print ( connection.get_dsn_parameters(),"\n") #create table Service Platforms get_project_name = "SELECT project_name FROM service_platforms WHERE name=\'" +self.name+ "\'" print (get_project_name) cursor.execute(get_project_name) all = cursor.fetchall() #return jsonify(all), 200 type_0 = all.__str__() print(type_0) type_1 = type_0[3:] print(type_1) type_2 = type_1[:-4] print(type_2) return type_2 except (Exception, psycopg2.Error) as error : print (error) exception_message = str(error) return exception_message, 401 finally: #closing database connection. if(connection): cursor.close() connection.close() print("PostgreSQL connection is closed") def getDBPassword(self): try: db = database.Database(FILE) connection = psycopg2.connect(user = db.user, password = db.password, host = db.host, port = db.port, database = db.database) cursor = connection.cursor() print ( connection.get_dsn_parameters(),"\n") #create table Service Platforms get_password= "SELECT username FROM service_platforms WHERE name=\'" +self.name+ "\'" print (get_password) cursor.execute(get_password) all = cursor.fetchall() #return jsonify(all), 200 type_0 = all.__str__() print(type_0) type_1 = type_0[3:] print(type_1) type_2 = type_1[:-4] print(type_2) return type_2 except (Exception, psycopg2.Error) as error : print (error) exception_message = str(error) return exception_message, 401 finally: #closing database connection. if(connection): cursor.close() connection.close() print("PostgreSQL connection is closed") def getDBProject(self): try: db = database.Database(FILE) connection = psycopg2.connect(user = db.user, password = db.password, host = db.host, port = db.port, database = db.database) cursor = connection.cursor() print ( connection.get_dsn_parameters(),"\n") #create table Service Platforms get_password= "SELECT project_name FROM service_platforms WHERE name=\'" +self.name+ "\'" print (get_password) cursor.execute(get_password) all = cursor.fetchall() #return jsonify(all), 200 type_0 = all.__str__() print(type_0) type_1 = type_0[3:] print(type_1) type_2 = type_1[:-4] print(type_2) return type_2 except (Exception, psycopg2.Error) as error : print (error) exception_message = str(error) return exception_message, 401 finally: #closing database connection. if(connection): cursor.close() connection.close() print("PostgreSQL connection is closed") def getDBHost(self): try: db = database.Database(FILE) connection = psycopg2.connect(user = db.user, password = db.password, host = db.host, port = db.port, database = db.database) cursor = connection.cursor() print ( connection.get_dsn_parameters(),"\n") #create table Service Platforms print (self.name) get_host = "SELECT host FROM service_platforms WHERE name=\'" +self.name+ "\'" print (get_host) cursor.execute(get_host) all = cursor.fetchall() #return jsonify(all), 200 return all, 200 except (Exception, psycopg2.Error) as error : print (error) exception_message = str(error) return exception_message, 401 finally: #closing database connection. if(connection): cursor.close() connection.close() print("PostgreSQL connection is closed") def getPackages(self): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/packages' #url = sp_url + '/packages' response = requests.get(url, headers=JSON_CONTENT_HEADER) if response.ok: return (response.text, response.status_code, response.headers.items()) if my_type == 'osm': return "osm packages" def getPackage(self,name,vendor,version): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/packages' print (name,vendor,version) response = requests.get(url,headers=JSON_CONTENT_HEADER) response_json = response.content jjson = json.loads(response_json) pkg = [x for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version] if response.ok: print(pkg) return jsonify(pkg) def deletePackage(self,name,vendor,version): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/packages' print (name,vendor,version) response = requests.get(url,headers=JSON_CONTENT_HEADER) response_json = response.content jjson = json.loads(response_json) pkg = [x for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version] if pkg: print(pkg) #uuid_to_delete = pkg['pd']['uuid'] #uuid_to_delete_1 = [uuid for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version] uuid_to_delete_1 = [obj['uuid'] for obj in jjson if(obj['pd']['name'] == name)] print(uuid_to_delete_1) uuid_0 = uuid_to_delete_1.__str__() uuid_to_delete_2 = uuid_0[2:] print(uuid_to_delete_2) uuid_to_delete_3 = uuid_to_delete_2[:-2] print(uuid_to_delete_3) url_for_delete = url + '/' + uuid_to_delete_3 print (url_for_delete) delete = requests.delete(url_for_delete, headers=JSON_CONTENT_HEADER) if response.ok: return (delete.text, delete.status_code, delete.headers.items()) def getPackagebyId(self,name,vendor,version): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/packages' print (name,vendor,version) response = requests.get(url,headers=JSON_CONTENT_HEADER) response_json = response.content jjson = json.loads(response_json) pkg = [x for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version] if pkg: print(pkg) #uuid_to_delete = pkg['pd']['uuid'] #uuid_to_delete_1 = [uuid for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version] uuid_to_delete_1 = [obj['uuid'] for obj in jjson if(obj['pd']['name'] == name)] print(uuid_to_delete_1) uuid_0 = uuid_to_delete_1.__str__() uuid_to_delete_2 = uuid_0[2:] print(uuid_to_delete_2) uuid_to_delete_3 = uuid_to_delete_2[:-2] print(uuid_to_delete_3) url_for_delete = url + '/' + uuid_to_delete_3 print (url_for_delete) delete = requests.get(url_for_delete, headers=JSON_CONTENT_HEADER) if response.ok: return (delete.text, delete.status_code, delete.headers.items()) def uploadPackage(self,package): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/packages' print(package) print(url) files = {'package': open(package,'rb')} upload = requests.post(url, files=files) if request.method == 'POST': return upload.text #curl -v -i -X POST -F "package=@./eu.5gtango.test-connectivity.0.1.tgo" http://sta-vnv-ave-v4-0.5gtango.eu:32002/api/v3/packages #upload_nsd_3 = upload_nsd_2 + " --data-binary " #upload_nsd_4 = upload_nsd_3 + "\"@" +file_to_upload+ "\" " + url_2 #print (upload_nsd_4) #upload = subprocess.check_output([upload_nsd_4], shell=True) #return jsonify(upload_nsd_4) #return (upload) #upload_package_1="curl -v -i -X POST -F \"package=@." + package + "\" " + url #return upload_package_1 def uploadOSMService(self,request): #JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'osm': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) content = request.get_json() print ("request content:") print (content) token = self.getOSMToken(request) print (token) file_to_upload = content['service'] print ("File to upload") print (file_to_upload) file_composed = "@" + file_to_upload print (file_composed) #file = {'file': open(file_to_upload, 'rb')} file = {'nsd-create': open(file_to_upload, 'rb')} print (file) data = {'service':file_to_upload} #data = request.get_json() print (data) HEADERS = { 'Accept':'application/yaml', 'Content-Type':'application/zip', 'Authorization':'Bearer ' +token+'' } print (HEADERS) url = sp_host_2 + ':9999/osm/nsd/v1/ns_descriptors_content' url_2 = url.replace("http","https") print (url_2) #upload_nsddddd = "osm --hostname " + sp_host_3 + " vnfd-create " + url_2 upload_nsd = "curl -X POST --insecure -w \"%{http_code}\" -H \"Content-type: application/zip\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer " upload_nsd_2 = upload_nsd +token + "\" " upload_nsd_3 = upload_nsd_2 + " --data-binary " upload_nsd_4 = upload_nsd_3 + "\"@" +file_to_upload+ "\" " + url_2 print (upload_nsd_4) upload = subprocess.check_output([upload_nsd_4], shell=True) #return jsonify(upload_nsd_4) return (upload) def uploadOSMFunction(self,request): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'osm': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) token = self.getOSMToken(request) print (token) content = request.get_json() file_to_upload = content['function'] url = sp_host_2 + ':9999/osm/vnfpkgm/v1/vnf_packages_content' url_2 = url.replace("http","https") print (url_2) upload_nsd = "curl -X POST --insecure -w \"%{http_code}\" -H \"Content-type: application/zip\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer " upload_nsd_2 = upload_nsd +token + "\" " upload_nsd_3 = upload_nsd_2 + " --data-binary " upload_nsd_4 = upload_nsd_3 + "\"@" +file_to_upload+ "\" " + url_2 print (upload_nsd_4) upload = subprocess.check_output([upload_nsd_4], shell=True) #return jsonify(upload_nsd_4) return (upload) def getServices(self): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() print (my_type) if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/services' print (url) #url = sp_url + '/packages' response = requests.get(url, headers=JSON_CONTENT_HEADER) if response.ok: return (response.text, response.status_code, response.headers.items()) if my_type == 'osm': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_2 + ':32002/api/v3/services' get_nsd_list = "osm --hostname " + sp_host_3 + " nsd-list" print (get_nsd_list) #get = os.system(get_nsd_list).__str__() #return get get = subprocess.check_output([get_nsd_list], shell=True) return (get) def getFunctions(self): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/functions' #url = sp_url + '/packages' response = requests.get(url, headers=JSON_CONTENT_HEADER) if response.ok: return (response.text, response.status_code, response.headers.items()) if my_type == 'osm': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_2 + ':32002/api/v3/functions' get_vnfd_list = "osm --hostname " + sp_host_3 + " vnfd-list" print (get_vnfd_list) #get = os.system(get_nsd_list).__str__() #return get get = subprocess.check_output([get_vnfd_list], shell=True) return (get) def getService(self,name,vendor,version): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/services' print (name,vendor,version) response = requests.get(url,headers=JSON_CONTENT_HEADER) response_json = response.content jjson = json.loads(response_json) pkg = [x for x in jjson if x['nsd']['name'] == name and x['nsd']['vendor'] == vendor and x['nsd']['version'] == version] if response.ok: print(pkg) return jsonify(pkg) def getServiceInstantiations(self,name,vendor,version): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/requests' print (name,vendor,version) response = requests.get(url,headers=JSON_CONTENT_HEADER) response_json = response.content #response_json = response.text print (response_json) #json_str = json.dumps(response._content) #jjson = json.loads(response.text.__str__()) #jjson = json.loads(response_json) jjson = json.loads(response.content) print (jjson) #services = [x for x in jjson if x['service']['name'] == name and x['service']['vendor'] == vendor and x['service']['version'] == version] #print (jjson['service']['name']) #print (obj for obj in jjson if(obj['service']['vendor'] == 'eu.5gtango')) #illo = [obj for obj in jjson if(obj['service'] == '11111')] print ("illo") #idd = print ([obj for obj in jjson[0]['service']['uuid']]) idd = print (jjson[0]['service']['uuid']) idd = print (jjson[0]['service']['name']) idd = print (jjson[1]['service']['uuid']) idd = print (jjson[1]['service']['name']) #idd = [obj.service['name'] for obj in jjson] print ("illo") N = 0 for N in range(10000): print (jjson['service']['uuid']) N = N + 1 print (N) if response.ok: #return (response.text, response.status_code, response.headers.items()) return jsonify("no") def getServiceId(self,name,vendor,version): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/services' print (name,vendor,version) response = requests.get(url,headers=JSON_CONTENT_HEADER) response_json = response.content print (response_json) jjson = json.loads(response_json) pkg = [x for x in jjson if x['nsd']['name'] == name and x['nsd']['vendor'] == vendor and x['nsd']['version'] == version] if pkg: print(pkg) uuid_to_delete_1 = [obj['uuid'] for obj in jjson if(obj['nsd']['name'] == name)] print(uuid_to_delete_1) uuid_0 = uuid_to_delete_1.__str__() uuid_to_delete_2 = uuid_0[2:] print(uuid_to_delete_2) uuid_to_delete_3 = uuid_to_delete_2[:-2] print(uuid_to_delete_3) url_for_delete = url + '/' + uuid_to_delete_3 print (url_for_delete) delete = requests.get(url_for_delete, headers=JSON_CONTENT_HEADER) if response.ok: return uuid_to_delete_3 def instantiationStatus(self,request): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': print('this SP is a Sonata') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) #url = sp_host_2 + ':32002/api/v3/requests/' + id url = sp_host_2 + ':32002/api/v3/requests/' + request print (url) response = requests.get(url,headers=JSON_CONTENT_HEADER) response_json = response.content print (response_json) #return response_json if response.ok: return (response.text, response.status_code, response.headers.items()) if my_type == 'osm': print('this SP is a OSM') print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 token = self.getOSMToken(request) print (token) content = request.get_json() ns_id = content['ns_id'] print (ns_id) url = sp_host_2 + ':9999/osm/nslcm/v1/ns_instances' url_2 = url.replace("http","https") print (url_2) status_ns = "curl --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer " status_ns_2 = status_ns +token + "\" " status_ns_3 = status_ns_2 + " " + url_2 + "/" + ns_id print (status_ns_3) status = subprocess.check_output([status_ns_3], shell=True) return (status) def instantiationsStatus(self): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': print('this SP is a Sonata') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/requests' print (url) response = requests.get(url,headers=JSON_CONTENT_HEADER) response_json = response.content print (response_json) #return response_json if response.ok: return (response.text, response.status_code, response.headers.items()) #print("status") #return "status" if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 get_status = "osm --hostname " + sp_host_3 + " ns-list" print (get_status) status = subprocess.check_output([get_status], shell=True) return (status) def instantiation(self,request): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} JSON_CONTENT_HEADER = {'content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': print('this SP is a Sonata') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/requests' print (url) print(request.get_json()) data = request.get_json() print(url) print (data) #upload = requests.post(url, files=files) #upload = requests.post(url, files=files) #instantiate = requests.post(url,data,headers=JSON_CONTENT_HEADER) instantiate = requests.post( url, data=json.dumps(data), headers=JSON_CONTENT_HEADER) print (instantiate) if request.method == 'POST': return instantiate.text if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 print(request.get_json()) content = request.get_json() print(content) token = self.getOSMToken(request) print (token) content = request.get_json() print (content) print (content['nsd_name']) print (content['ns_name']) print (content['vim_account']) url = sp_host_2 + ':9999/osm/nslcm/v1/ns_instances_content' url_2 = url.replace("http","https") print (url_2) print (content['vim_account']) vim_id = self.getVimId(content['vim_account']) print (vim_id) print (content['nsd_name']) nsd_id = self.getOSMNsdId(content['nsd_name']) print (nsd_id) HEADERS = { 'Accept':'application/json', 'Content-Type':'application/json', 'Authorization':'Bearer ' +token+'' } data_inst = { 'nsdId':''+nsd_id+'', 'nsName':''+content['ns_name']+'', 'vimAccountId':''+vim_id+'' } instantiate_nsd = "curl -X POST --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer " instantiate_nsd_2 = instantiate_nsd +token + "\" " instantiate_nsd_3 = instantiate_nsd_2 + " --data \"" + str(data_inst) + "\"" instantiate_nsd_4 = instantiate_nsd_3 + " " + url_2 print (instantiate_nsd_4) inst = subprocess.check_output([instantiate_nsd_4], shell=True) return (inst) def instantiationDelete(self,request): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': print('this SP is a Sonata') print('this SP is a Sonata') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/requests' print(request.get_json()) data = request.get_json() print(url) #upload = requests.post(url, files=files) #upload = requests.post(url, files=files) terminate = requests.post(url,data,headers=JSON_CONTENT_HEADER) if request.method == 'POST': return terminate.text if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 print(request.get_json()) print(url) token = self.getOSMToken(request) print (token) content = request.get_json() url = sp_host_2 + ':9999/osm/nslcm/v1/ns_instances_content' url_2 = url.replace("http","https") print (url_2) print (content['ns_id']) ns_id = content['ns_id'] print (ns_id) delete_ns = "curl -X DELETE --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer " delete_ns_2 = delete_ns +token + "\" " delete_ns_3 = delete_ns_2 + " " + url_2 + "/" + ns_id print (delete_ns_3) terminate = subprocess.check_output([delete_ns_3], shell=True) return (terminate) #delete_ns = "osm --hostname " + sp_host_3 + " ns-delete " + data['ns_name'] #print (delete_ns) #delete = subprocess.check_output([delete_ns], shell=True) #return (delete) def getOSMToken(self,request): #JSON_CONTENT_HEADER = {'Content-Type':'application/json'} JSON_CONTENT_HEADER = {'Accept':'application/json'} my_type = self.getDBType() if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) #url = sp_host_2 + '/requests' url = sp_host_2 + ':9999/osm/admin/v1/tokens' url_2 = url.replace("http","https") print (url_2) print(request.get_json()) #data = request.get_json() print(url_2) #print (data) #print (data['nsd_name']) #print (data['username']) #print (data['password']) #print (data['project_id']) pr_name = self.getDBProjectName() print ("project name from DB:") print (pr_name) if pr_name: project_id_for_token = pr_name if not pr_name: data = request.get_json() project_id_for_token = data['project_id'] print ("project name from json body:") print (pr_name) #project_id_for_token = data['project_id'] print (project_id_for_token) #username_for_token = data['username'] #password_for_token = data['password'] username_for_token = self.getDBUserName() password_for_token = self.getDBPassword() #project_id_for_token = data['project_id'] admin_data = "{username: 'admin', password: 'admin', project_id: 'admin'}" print (admin_data) #update_token = "UPDATE service_platforms SET service_token = \'" +token+ "\' WHERE name = \'" +self.name+ "\'" data_for_token= "{username: \'" +username_for_token+ "\', password: \'" +password_for_token+ "\', project_id: \'" +project_id_for_token+ "\'}" #print (data) get_token = requests.post(url_2,data=data_for_token,headers=JSON_CONTENT_HEADER,verify=False) print (get_token.text) print (get_token.content) token_id = get_token.json() print (token_id['id']) #upd_tok = self.updateToken(token_id['id']) #print (upd_tok) return token_id['id'] def getVims(self): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': print('this SP is a Sonata') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/requests' print (url) return url if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 get_vims = "osm --hostname " + sp_host_3 + " vim-list" print (get_vims) vims = subprocess.check_output([get_vims], shell=True) return (vims) def getVim(self,vim): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': print('this SP is a Sonata') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/requests' print (url) return url if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 get_vim = "osm --hostname " + sp_host_3 + " vim-show " + vim print (get_vim) vim_info = subprocess.check_output([get_vim], shell=True) return jsonify(vim_info) def getVimId(self,vim): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': print('this SP is a Sonata') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/requests' print (url) return url if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 get_vim = "osm --hostname " + sp_host_3 + " vim-show " + vim print (get_vim) vim_info = subprocess.check_output([get_vim], shell=True) print (vim_info) print (type(vim_info)) s = json.dumps(str(vim_info)) print(s) print (type(s)) print ("ILLO") start = s.find('_id') end = s.find('\\\" ', start) print (s[start+20:end]) vim_id = s[start+20:end] print ("ILLO") return vim_id def getOSMNsdId(self,nsd_name): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 get_nsd = "osm --hostname " + sp_host_3 + " nsd-show " + nsd_name print (get_nsd) nsd_info = subprocess.check_output([get_nsd], shell=True) print (nsd_info) print (type(nsd_info)) s = json.dumps(str(nsd_info)) print(s) print (type(s)) print ("ILLO") start = s.find('_id') end = s.find('\\\" ', start) #print (s[start+21:end]) #vim_id = s[start+21:end] print (s[start+21:end]) vim_id = s[start+21:end] print ("ILLO") return vim_id def downloadPackage(self): JSON_CONTENT_HEADER = {'Content-Type':'application/json'} my_type = self.getDBType() if my_type == 'sonata': sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) url = sp_host_2 + ':32002/api/v3/packages' print(url) files = {'package': open(package,'rb')} upload = requests.post(url, files=files) if request.method == 'POST': return upload.text def deleteOSMService(self,id_to_delete): print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 token = self.getOSMTokenForDelete() print (token) url = sp_host_2 + ':9999/osm/nsd/v1/ns_descriptors_content' url_2 = url.replace("http","https") print (url_2) delete_nsd = "curl --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer " delete_nsd_2 = delete_nsd +token + "\" " + url_2 + "/" + id_to_delete + " -X DELETE" print (delete_nsd_2) deletion = subprocess.check_output([delete_nsd_2], shell=True) return (deletion) def deleteOSMFunction(self,id_to_delete): print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) sp_host_3 = sp_host_2[7:] print ("sp3 es: ") print (sp_host_3) url = sp_host_3 token = self.getOSMTokenForDelete() print (token) url = sp_host_2 + ':9999/osm/vnfpkgm/v1/vnf_packages' url_2 = url.replace("http","https") print (url_2) delete_nsd = "curl --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer " delete_nsd_2 = delete_nsd +token + "\" " + url_2 + "/" + id_to_delete + " -X DELETE" print (delete_nsd_2) deletion = subprocess.check_output([delete_nsd_2], shell=True) return (deletion) def getOSMTokenForDelete(self): #JSON_CONTENT_HEADER = {'Content-Type':'application/json'} JSON_CONTENT_HEADER = {'Accept':'application/json'} my_type = self.getDBType() if my_type == 'osm': print('this SP is a OSM') sp_host_0 = self.getDBHost() print (sp_host_0) sp_host = sp_host_0.__str__() print (sp_host) #print (self.getDBHost()) sp_host_1 = sp_host[4:] print ("sp1 es: ") print (sp_host_1) sp_host_2 = sp_host_1[:-10] print ("sp2 es: ") print (sp_host_2) #url = sp_host_2 + '/requests' url = sp_host_2 + ':9999/osm/admin/v1/tokens' url_2 = url.replace("http","https") print (url_2) #data = request.get_json() print(url_2) #print (data) #print (data['nsd_name']) #print (data['username']) #print (data['password']) #print (data['project_id']) pr_name = self.getDBProjectName() print ("project name from DB:") print (pr_name) if pr_name: project_id_for_token = pr_name if not pr_name: #data = request.get_json() #project_id_for_token = data['project_id'] project_id_for_token = self.getDBProject(self) print ("project name from json body:") print (pr_name) #project_id_for_token = data['project_id'] print (project_id_for_token) #username_for_token = data['username'] #password_for_token = data['password'] username_for_token = self.getDBUserName() password_for_token = self.getDBPassword() #project_id_for_token = data['project_id'] admin_data = "{username: 'admin', password: 'admin', project_id: 'admin'}" print (admin_data) #update_token = "UPDATE service_platforms SET service_token = \'" +token+ "\' WHERE name = \'" +self.name+ "\'" data_for_token= "{username: \'" +username_for_token+ "\', password: \'" +password_for_token+ "\', project_id: \'" +project_id_for_token+ "\'}" #print (data) get_token = requests.post(url_2,data=data_for_token,headers=JSON_CONTENT_HEADER,verify=False) print (get_token.text) print (get_token.content) token_id = get_token.json() print (token_id['id']) #upd_tok = self.updateToken(token_id['id']) #print (upd_tok) return token_id['id']
34.288916
174
0.488333
6,216
56,611
4.148327
0.041828
0.108664
0.063988
0.04134
0.877608
0.856395
0.842938
0.836539
0.824556
0.8187
0
0.029934
0.402802
56,611
1,651
175
34.288916
0.732785
0.08622
0
0.842472
0
0
0.095298
0.014614
0
0
0
0
0
1
0.033943
false
0.017406
0.008703
0.002611
0.088773
0.382942
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
98a299fc11240dce08dded4ed8e707340fd1e252
21,372
py
Python
tests/all_tests.py
tjweisman/dixit
57933a78200a8738b973a021ec8d3c3b832d1f7a
[ "MIT" ]
null
null
null
tests/all_tests.py
tjweisman/dixit
57933a78200a8738b973a021ec8d3c3b832d1f7a
[ "MIT" ]
null
null
null
tests/all_tests.py
tjweisman/dixit
57933a78200a8738b973a021ec8d3c3b832d1f7a
[ "MIT" ]
null
null
null
from selenium import webdriver from selenium.webdriver.support.ui import WebDriverWait from selenium.webdriver.common.by import By import pages import subtests def login(): with webdriver.Firefox() as driver: driver.get(pages.DIXIT_URL) start_page = pages.StartPage(driver) start_page.delete_users() pregame = start_page.join_game("test_user", "test_room") return "test_user" in pregame.users_list() def login_multiple(): with webdriver.Firefox() as driver: users = ["test_1", "test_2"] session = pages.create_game_session(driver, users, "test_room") pregame = pages.PregamePage(session.driver) result_users = pregame.users_list() return all([user in result_users for user in users]) def two_player_single_round(): with webdriver.Firefox() as driver: users = ["test_1", "test_2"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() game_correct = subtests.play_all_correct_round(session, gameplay, [0, 1], users, 0, ["0", "2"]) return game_correct def two_player_double_round(): with webdriver.Firefox() as driver: users = ["test_1", "test_2"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() round1_correct = subtests.play_all_correct_round(session, gameplay, [0,1], users, 0, ["0", "2"]) round2_correct = subtests.play_all_correct_round(session, gameplay, [0,1], users, 1, ["2", "2"]) return round1_correct and round2_correct def three_player_single_round(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) p2_card = gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.click_cid(correct_cid) gameplay.submit_guess() session.activate(1) gameplay.click_cid(p2_card) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '0', '4']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_3'], [], ['test_2']]) return scores_correct and cards_correct and choices_correct def leave_before_prompt(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() session.activate(2) gameplay.leave_game() session.activate(0) gameplay.click_hand_card(0) gameplay.submit_hint() session.activate(1) gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.click_allowable_table_card() gameplay.submit_guess() users = ['test_1', 'test_2'] scores_correct = (gameplay.get_scores(users) == ['0', '2']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_2'], []]) return scores_correct and cards_correct and choices_correct def leave_after_prompt(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() session.activate(0) gameplay.click_hand_card(0) gameplay.submit_hint() session.activate(2) gameplay.leave_game() session.activate(1) gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.click_allowable_table_card() gameplay.submit_guess() users = ['test_1', 'test_2'] scores_correct = (gameplay.get_scores(users) == ['0', '2']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_2'], []]) return scores_correct and cards_correct and choices_correct def leave_after_secret(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.submit_hint() session.activate(2) gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.leave_game() session.activate(1) gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.click_cid(correct_cid) gameplay.submit_guess() remaining_users = ['test_1', 'test_2'] scores_correct = (gameplay.get_scores(remaining_users) == ['0', '2']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_2'], [], []]) return scores_correct and cards_correct and choices_correct def leave_before_guess(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.submit_hint() session.activate(2) gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(1) gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) gameplay.leave_game() session.activate(1) gameplay.click_cid(correct_cid) gameplay.submit_guess() remaining_users = ['test_1', 'test_2'] scores_correct = (gameplay.get_scores(remaining_users) == ['0', '2']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_2'], [], []]) return scores_correct and cards_correct and choices_correct def leave_after_guess(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.submit_hint() session.activate(2) p3_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(1) gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) gameplay.click_cid(correct_cid) gameplay.submit_guess() gameplay.leave_game() session.activate(1) gameplay.click_cid(p3_card) gameplay.submit_guess() remaining_users = ['test_1', 'test_2'] scores_correct = (gameplay.get_scores(remaining_users) == ['3', '0']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_3'], [], ['test_2']]) return scores_correct and cards_correct and choices_correct def refresh_before_prompt(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() startpage = session.login_page() session.activate(2) session.refresh() startpage.join_game("test_3", "test_room") return subtests.play_all_correct_round(session, gameplay, [0,1,2], users, 0, ["0", "2", "2"]) def refresh_after_prompt(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() startpage = session.login_page() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(2) session.refresh() startpage.join_game("test_3", "test_room") session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) p2_card = gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.click_cid(correct_cid) gameplay.submit_guess() session.activate(1) gameplay.click_cid(p2_card) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '0', '4']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_3'], [], ['test_2']]) return scores_correct and cards_correct and choices_correct def refresh_after_secret(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() startpage = session.login_page() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(2) p2_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.refresh() startpage.join_game("test_3", "test_room") session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) gameplay.click_cid(correct_cid) gameplay.submit_guess() session.activate(1) gameplay.click_cid(p2_card) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '0', '4']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_3'], [], ['test_2']]) return scores_correct and cards_correct and choices_correct def refresh_before_guess(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() startpage = session.login_page() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(2) p2_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) session.refresh() startpage.join_game("test_3", "test_room") gameplay.click_cid(correct_cid) gameplay.submit_guess() session.activate(1) gameplay.click_cid(p2_card) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '0', '4']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_3'], [], ['test_2']]) return scores_correct and cards_correct and choices_correct def refresh_after_guess(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") gameplay = session.start_game() startpage = session.login_page() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(2) p2_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(1) gameplay.click_cid(p2_card) gameplay.submit_guess() session.activate(2) session.refresh() startpage.join_game("test_3", "test_room") gameplay.click_cid(correct_cid) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '0', '4']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_3'], [], ['test_2']]) return scores_correct and cards_correct and choices_correct def rejoin_before_prompt(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") pregame = session.pregame_page() session.activate(2) pregame.leave_game() session.activate(0) gameplay = session.start_game() startpage = session.login_page() session.activate(2) startpage.join_game("test_3", "test_room") return subtests.play_all_correct_round(session, gameplay, [0,1,2], users, 0, ["0", "2", "2"]) def rejoin_after_prompt(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") pregame = session.pregame_page() session.activate(2) pregame.leave_game() session.activate(0) gameplay = session.start_game() startpage = session.login_page() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(2) startpage.join_game("test_3", "test_room") session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) p2_card = gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.click_cid(correct_cid) gameplay.submit_guess() session.activate(1) gameplay.click_cid(p2_card) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '0', '4']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_3'], [], ['test_2']]) return scores_correct and cards_correct and choices_correct def rejoin_after_secret(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3", "test_4"] session = pages.create_game_session(driver, users, "test_room") pregame = session.pregame_page() session.activate(3) pregame.leave_game() session.activate(0) gameplay = session.start_game() startpage = session.login_page() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(3) startpage.join_game("test_4", "test_room") p3_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) p2_card = gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.click_cid(correct_cid) gameplay.submit_guess() session.activate(1) gameplay.click_cid(p3_card) gameplay.submit_guess() session.activate(3) gameplay.click_cid(correct_cid) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '0', '3', '4']) cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(users) == [['test_3', 'test_4'], [], [], ['test_2']]) return scores_correct and cards_correct and choices_correct def rejoin_before_guess(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3"] session = pages.create_game_session(driver, users, "test_room") pregame = session.pregame_page() session.activate(2) pregame.leave_game() session.activate(0) gameplay = session.start_game() startpage = session.login_page() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) startpage.join_game("test_3", "test_room") gameplay.click_cid(p1_card) gameplay.submit_guess() session.activate(1) gameplay.click_cid(correct_cid) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '4', '0']) secret_users = ["test_1", "test_2"] cards_correct = (gameplay.get_card_states(secret_users) == ['correct-uid', 'inactive']) choices_correct = (gameplay.get_player_choices(secret_users) == [['test_2'], ['test_3']]) return scores_correct and cards_correct and choices_correct def rejoin_after_guess(): with webdriver.Firefox() as driver: users = ["test_1", "test_2", "test_3", "test_4"] session = pages.create_game_session(driver, users, "test_room") pregame = session.pregame_page() session.activate(3) pregame.leave_game() session.activate(0) gameplay = session.start_game() startpage = session.login_page() session.activate(0) correct_cid = gameplay.click_hand_card(0) gameplay.set_hint_text("test hint") gameplay.submit_hint() session.activate(1) p1_card = gameplay.click_hand_card(0) gameplay.submit_secret() session.activate(2) p2_card = gameplay.click_hand_card(0) gameplay.submit_secret() gameplay.click_cid(p1_card) gameplay.submit_guess() session.activate(3) startpage.join_game("test_4", "test_room") gameplay.click_cid(correct_cid) gameplay.submit_guess() session.activate(1) gameplay.click_cid(p2_card) gameplay.submit_guess() scores_correct = (gameplay.get_scores(users) == ['3', '1', '1', '3']) secret_users = ["test_1", "test_2", "test_3"] cards_correct = (gameplay.get_card_states(secret_users) == ['correct-uid', 'inactive', 'inactive']) choices_correct = (gameplay.get_player_choices(secret_users) == [['test_4'], ['test_3'], ['test_2']]) return scores_correct and cards_correct and choices_correct
33.709779
88
0.588106
2,351
21,372
5.041684
0.037006
0.094913
0.063781
0.070868
0.955032
0.95267
0.950477
0.947355
0.946174
0.94474
0
0.021561
0.296884
21,372
633
89
33.763033
0.767219
0
0
0.89852
0
0
0.066629
0
0
0
0
0
0
1
0.042283
false
0
0.010571
0
0.095137
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
7f2d580dba7c737f6ab1c63302ec61219ac2dd47
8,965
py
Python
luiz_viana/graficos_luiz_viana.py
lfelipeviana/visualizacao_covid
9e061e2260c585e84956225d090838fdb3ee9e82
[ "MIT" ]
null
null
null
luiz_viana/graficos_luiz_viana.py
lfelipeviana/visualizacao_covid
9e061e2260c585e84956225d090838fdb3ee9e82
[ "MIT" ]
null
null
null
luiz_viana/graficos_luiz_viana.py
lfelipeviana/visualizacao_covid
9e061e2260c585e84956225d090838fdb3ee9e82
[ "MIT" ]
null
null
null
import dash_core_components as dcc import dash_html_components as html from dash.dependencies import Input, Output import pandas as pd from .barras_fx_etaria import barras_fx_etaria, barras_sub_plot_vacina def grafico_total_vacinas_idade(app): df = pd.read_csv("luiz_viana/data_set_consolidado_vacina.csv") df = df[df['paciente_endereco_uf'] != 'XX'] estados = df['paciente_endereco_uf'].unique().tolist() estados.sort() estados = ['Todos'] + estados opcoes = [{'label': i, 'value': i} for i in estados] sexo = ['Masculino', 'Feminino'] opcoes_sexo = [{'label': i, 'value': i[0]} for i in sexo] fx_etaria = df['faixa_etaria'].unique().tolist() fx_etaria = [int(item) for item in fx_etaria] fx_etaria.sort() opcoes_fx_etaria = [{'label': 'de {} a {}'.format(i, i+10 ), 'value': i} for i in fx_etaria] component_html = html.Div([ html.Hr(), html.H4('Vacinas aplicadas por dia e faixa etária'), html.Label("Filtros"), html.Div( [ html.Div([ html.Label("Sexo:"), html.Label("Faixa Etária:"), #html.Label("Estado"), ], style={'width': '10%', 'display': 'inline-block'}), html.Div( [ dcc.Checklist( id='sexo-filter', options=opcoes_sexo, value=['M','F'], labelStyle={'display': 'inline-block'} ), dcc.Checklist( id='fx_etaria-filter', options=opcoes_fx_etaria, value=fx_etaria, labelStyle={'display': 'inline-block'} ), ], style={'width': '90%', 'display': 'inline-block'}), ], className="row"), html.Div( [ html.Div( [ html.Label("Estado"), ], style={'width': '10%', 'display': 'inline-block'}), html.Div([ dcc.Dropdown( id='estado-filter', options=opcoes, value='Todos', style={"width": "40%"} ), ], style={'width': '90%', 'display': 'inline-block'}), ]), html.Div( [ dcc.Graph(id='vacinas-fx_etaria') ]) ]) @app.callback( Output('vacinas-fx_etaria', 'figure'), Input('estado-filter', 'value'), Input('sexo-filter', 'value'), Input('fx_etaria-filter', 'value')) def update_dataset_fx_etaria(estado, sexo, fx_etaria): df = pd.read_csv("luiz_viana/data_set_consolidado_vacina.csv") df = df[df['paciente_endereco_uf'] != 'XX'] r_min = 0 r_max = df.groupby(['vacina_dataaplicacao'], as_index=False)['quantidade'].sum()['quantidade'].max() + 100000 fx_etaria_opts = df['faixa_etaria'].unique().tolist() if estado != "Todos": df = df[df['paciente_endereco_uf'] == estado] r_min = 0 r_max = df.groupby(['vacina_dataaplicacao'], as_index=False)['quantidade'].sum()['quantidade'].max() + 100000 for sx in ('M','F'): if sx not in sexo: df = df[df['paciente_enumsexobiologico'] != sx] if fx_etaria is not None: for fx in fx_etaria_opts: if fx not in fx_etaria: df = df[df['faixa_etaria'] != fx] data_fx = df.groupby(['vacina_dataaplicacao', 'faixa_etaria'], as_index=False)['quantidade'].sum() fx0 = [{'vacina_dataaplicacao': '2021-01-31', 'faixa_etaria': 0, 'quantidade': 0}] fx100 = [{'vacina_dataaplicacao': '2021-01-31', 'faixa_etaria': 100, 'quantidade': 0}] data_fx.loc[len(data_fx.index)]=list(fx0[0].values()) data_fx.loc[len(data_fx.index)]=list(fx100[0].values()) fig = barras_fx_etaria(data_fx, r_min, r_max) return fig return component_html def grafico_sub_plot_vacinas_idade(app): df = pd.read_csv("luiz_viana/data_set_consolidado_vacina.csv") df = df[df['paciente_endereco_uf'] != 'XX'] estados = df['paciente_endereco_uf'].unique().tolist() estados.sort() estados = ['Todos'] + estados opcoes = [{'label': i, 'value': i} for i in estados] sexo = ['Masculino', 'Feminino'] opcoes_sexo = [{'label': i, 'value': i[0]} for i in sexo] fx_etaria = df['faixa_etaria'].unique().tolist() fx_etaria = [int(item) for item in fx_etaria] fx_etaria.sort() opcoes_fx_etaria = [{'label': 'de {} a {}'.format(i, i+10 ), 'value': i} for i in fx_etaria] component_html = html.Div([ html.Hr(), html.H4('Vacinas aplicadas por dia e faixa etária'), html.Label("Filtros"), html.Div( [ html.Div([ html.Label("Sexo:"), html.Label("Faixa Etária:"), #html.Label("Estado"), ], style={'width': '10%', 'display': 'inline-block'}), html.Div( [ dcc.Checklist( id='sexo-filter', options=opcoes_sexo, value=['M','F'], labelStyle={'display': 'inline-block'} ), dcc.Checklist( id='fx_etaria-filter', options=opcoes_fx_etaria, value=fx_etaria, labelStyle={'display': 'inline-block'} ), ], style={'width': '90%', 'display': 'inline-block'}), ], className="row"), html.Div( [ html.Div( [ html.Label("Estado"), ], style={'width': '10%', 'display': 'inline-block'}), html.Div([ dcc.Dropdown( id='estado-filter', options=opcoes, value='Todos', style={"width": "40%"} ), ], style={'width': '90%', 'display': 'inline-block'}), ]), html.Div( [ dcc.Graph(id='vacinas-fx_etaria') ]) ]) @app.callback( Output('vacinas-fx_etaria', 'figure'), Input('estado-filter', 'value'), Input('sexo-filter', 'value'), Input('fx_etaria-filter', 'value')) def update_dataset_fx_etaria(estado, sexo, fx_etaria): df = pd.read_csv("luiz_viana/data_set_consolidado_vacina.csv") df = df[df['paciente_endereco_uf'] != 'XX'] r_min = 0 r_max = df.groupby(['vacina_dataaplicacao'], as_index=False)['quantidade'].sum()['quantidade'].max() * 1.1 fx_etaria_opts = df['faixa_etaria'].unique().tolist() if estado != "Todos": df = df[df['paciente_endereco_uf'] == estado] r_min = 0 r_max = df.groupby(['vacina_dataaplicacao'], as_index=False)['quantidade'].sum()['quantidade'].max() * 1.1 for sx in ('M','F'): if sx not in sexo: df = df[df['paciente_enumsexobiologico'] != sx] if fx_etaria is not None: for fx in fx_etaria_opts: if fx not in fx_etaria: df = df[df['faixa_etaria'] != fx] data_fx = df.groupby(['vacina_dataaplicacao', 'faixa_etaria', 'vacina_nome'], as_index=False)['quantidade'].sum() fig = barras_sub_plot_vacina(data_fx, r_min, r_max) return fig return component_html def graficos_luiz_viana(app): components_html = [] #components_html.append(grafico_total_vacinas_idade(app)) components_html.append(grafico_sub_plot_vacinas_idade(app)) return components_html
34.348659
78
0.452426
877
8,965
4.425314
0.142531
0.080392
0.055656
0.028859
0.904406
0.881474
0.870394
0.850812
0.836898
0.836898
0
0.015198
0.412828
8,965
261
79
34.348659
0.722074
0.010931
0
0.845361
0
0
0.191427
0.024817
0
0
0
0
0
1
0.025773
false
0
0.025773
0
0.07732
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
7f488b8e381d84ced95304a49f4c70dda2aeb848
8,165
py
Python
PostDiffMixture/simulations_folder/simulation_analysis_scripts/phi_functions.py
SIGKDDanon/SIGKDD2021DeAnonV2
76f0373ec42ab55feefed3f4ce4bf4d532b51dd2
[ "Apache-2.0" ]
null
null
null
PostDiffMixture/simulations_folder/simulation_analysis_scripts/phi_functions.py
SIGKDDanon/SIGKDD2021DeAnonV2
76f0373ec42ab55feefed3f4ce4bf4d532b51dd2
[ "Apache-2.0" ]
null
null
null
PostDiffMixture/simulations_folder/simulation_analysis_scripts/phi_functions.py
SIGKDDanon/SIGKDD2021DeAnonV2
76f0373ec42ab55feefed3f4ce4bf4d532b51dd2
[ "Apache-2.0" ]
null
null
null
import ipdb as pb import pandas as pd import numpy as np from scipy.stats import beta CUML_KEY = "prop_exploring_ppd_cuml" SNAPSHOT_KKEY = "exploring_ppd_at_this_n" def plot_phi(df, num_sims, n, c, ax, es = 0): """ get prop in cond 1 for when exploring """ # pb.set_trace() step_sizes = [int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))] extra_steps = [3,4,5] phi_init = np.sum(np.abs(beta.rvs(1, 1, size=5000) - beta.rvs(1, 1, size=5000)) < c)/5000 phi_list = [phi_init] #per step sizes for num_steps in step_sizes:#loop over step sizes, make a hist for each print("num_steps", num_steps) # df[df["num_steps"] == 16]["exploring_ppd_at_this_n"].mean() df_for_num_steps = df[df['num_steps'] == num_steps].dropna() phi = df_for_num_steps[SNAPSHOT_KKEY].mean() phi_list.append(phi) # pb.set_trace() phi_list = np.array(phi_list) se = np.sqrt(phi_list*(1-phi_list))/np.sqrt(num_sims) # h = se * stats.t.ppf((1 + 0.95) / 2., num_sims-1) # h = stats.t.ppf(1-0.025, num_sims)*np.sqrt(phi_list*(1-phi_list)/num_sims) #95 CI for Proportion #print(phi_list, num_sims) step_sizes = [0, int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))] if c == 0.025: #0.05 phi_list_0 = [0, 0, 0, 0, 0] phi_list_0 = np.array(phi_list_0) ax.errorbar(step_sizes, phi_list_0,yerr = None,fmt = ".-", label = "c = 0") ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c)) # else: # ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c)) ind = np.arange(len(step_sizes)) ind = step_sizes ax.set_xticks(ind) print(step_sizes) ax.set_xticklabels(step_sizes) ax.tick_params(axis='x', rotation=45) eps_dec = step_sizes eps_dec[0] = eps_dec[0] + 1 print("EPS DEC") print(1.0/np.array(eps_dec)) eps_dec = np.arange(step_sizes[-1]) ax.errorbar(eps_dec, 1.0/np.sqrt(np.array(eps_dec)), color='black', linestyle='--', linewidth = 0.5) #ax.tick_params(axis='x') # ax.set_xlabel("number of participants = \n n/2, n, 2*n, 4*n") #if es == 0.1: ax.set_xlabel("number of participants") ax.set_ylim(-0.05, 1.05) if es == 0: ax.set_ylabel("$\hat \phi_t$") if es != 0: handles, labels = ax.get_legend_handles_labels() #ax.legend(handles[::-1], labels[::-1], loc='upper left') ax.legend(handles[::-1], labels[::-1], bbox_to_anchor=(1.1, 1.03)) # ax.legend(handles[::-1], labels[::-1]) #leg2 = ax.legend(loc = 2) def plot_phi_www(df, num_sims, n, c, ax = None, es = 0): """ get prop in cond 1 for when exploring """ # pb.set_trace() step_sizes = [int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))] #step_sizes = [int(np.ceil(n/2)), int(n)] extra_steps = [3,4,5] phi_init = np.sum(np.abs(beta.rvs(1, 1, size=5000) - beta.rvs(1, 1, size=5000)) < c)/5000 phi_list = [phi_init] #per step sizes # if c != 0: for num_steps in step_sizes:#loop over step sizes, make a hist for each print("num_steps", num_steps) # df[df["num_steps"] == 16]["exploring_ppd_at_this_n"].mean() df_for_num_steps = df[df['num_steps'] == num_steps].dropna() phi = df_for_num_steps[SNAPSHOT_KKEY].mean() phi_list.append(phi) # pb.set_trace() phi_list = np.array(phi_list) se = np.sqrt(phi_list*(1-phi_list))/np.sqrt(num_sims) # h = se * stats.t.ppf((1 + 0.95) / 2., num_sims-1) # h = stats.t.ppf(1-0.025, num_sims)*np.sqrt(phi_list*(1-phi_list)/num_sims) #95 CI for Proportion #print(phi_list, num_sims) step_sizes = [0, int(np.ceil(n/2)), int(n)] step_sizes = [0, int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))] ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c)) # else: # ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c)) ind = np.arange(len(step_sizes)) ind = step_sizes ax.set_xticks(ind) print(step_sizes) ax.set_xticklabels(step_sizes) #ax.tick_params(axis='x', rotation=45) ax.tick_params(axis='x') # ax.set_xlabel("number of participants = \n n/2, n, 2*n, 4*n") ax.set_xlabel("number of participants") ax.set_ylabel("$\hat \phi$") ax.set_ylim(-0.05, 1.05) print('HERHEHREHRHERHEHR') eps_dec = step_sizes eps_dec[0] = eps_dec[0] + 1 ax.errorbar(step_sizes, 1.0/np.array(eps_dec), color='black', linestyle='--') if c == 0.1 or 1: print("PHI 0") phi_list_0 = [0, 0, 0, 0, 0] phi_list_0 = np.array(phi_list_0) ax.errorbar(step_sizes, phi_list_0,yerr = None,fmt = ".-", label = "c = 0") if n == 785: handles, labels = ax.get_legend_handles_labels() #ax.legend(handles[::-1], labels[::-1], loc='upper left') ax.legend(handles[::-1], labels[::-1], bbox_to_anchor=(1.1, 1.05)) #leg2 = ax.legend(loc = 2) # fig, ax = plt.subplots(1,4, figsize = (12,5))#for phi # fig.tight_layout(rect=[0, 0.03, 1, 0.87]) # save_dir = "../simulation_analysis_saves/phi_plots_WWW/NoEffect/" # Path(save_dir).mkdir(parents=True, exist_ok=True) # # title = "$\hat \phi$ Across {} Simulations $p_1 = p_2 = 0.5$ \n $\phi$ := p($|p_1 - p_2| < c$) \n = {}".format(num_sims, n) # # fig.suptitle(title) # fig.savefig(save_dir + "/" + title +".png") def plot_phi_www_multi(df, df_ne, num_sims, n, c, ax = None, es = 0): """ get prop in cond 1 for when exploring """ # pb.set_trace() step_sizes = [int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))] #step_sizes = [int(np.ceil(n/2)), int(n)] extra_steps = [3,4,5] phi_init = np.sum(np.abs(beta.rvs(1, 1, size=5000) - beta.rvs(1, 1, size=5000)) < c)/5000 phi_list = [phi_init] #per step sizes for num_steps in step_sizes:#loop over step sizes, make a hist for each print("num_steps", num_steps) # df[df["num_steps"] == 16]["exploring_ppd_at_this_n"].mean() df_for_num_steps = df[df['num_steps'] == num_steps].dropna() phi = df_for_num_steps[SNAPSHOT_KKEY].mean() phi_list.append(phi) # pb.set_trace() phi_list = np.array(phi_list) se = np.sqrt(phi_list*(1-phi_list))/np.sqrt(num_sims) # h = se * stats.t.ppf((1 + 0.95) / 2., num_sims-1) # h = stats.t.ppf(1-0.025, num_sims)*np.sqrt(phi_list*(1-phi_list)/num_sims) #95 CI for Proportion #print(phi_list, num_sims) step_sizes = [0, int(np.ceil(n/2)), int(n)] step_sizes = [0, int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))] ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c)) # else: # ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c)) ind = np.arange(len(step_sizes)) ind = step_sizes ax.set_xticks(ind) print(step_sizes) ax.set_xticklabels(step_sizes) #ax.tick_params(axis='x', rotation=45) ax.tick_params(axis='x') # ax.set_xlabel("number of participants = \n n/2, n, 2*n, 4*n") ax.set_xlabel("number of participants") ax.set_ylabel("$\hat \phi$") #ax.set_ylim(0.0, 1.0) ax.set_ylim(-0.05, 1.05) if n == 785: handles, labels = ax.get_legend_handles_labels() #ax.legend(handles[::-1], labels[::-1], loc='upper left') ax.legend(handles[::-1], labels[::-1], bbox_to_anchor=(1.1, 1.05)) #leg2 = ax.legend(loc = 2) # fig, ax = plt.subplots(1,4, figsize = (12,5))#for phi # fig.tight_layout(rect=[0, 0.03, 1, 0.87]) # save_dir = "../simulation_analysis_saves/phi_plots_WWW/NoEffect/" # Path(save_dir).mkdir(parents=True, exist_ok=True) # # title = "$\hat \phi$ Across {} Simulations $p_1 = p_2 = 0.5$ \n $\phi$ := p($|p_1 - p_2| < c$) \n = {}".format(num_sims, n) # # fig.suptitle(title) # fig.savefig(save_dir + "/" + title +".png")
39.066986
132
0.589957
1,396
8,165
3.260029
0.113181
0.085036
0.043507
0.026368
0.929246
0.91782
0.905735
0.890574
0.87673
0.87673
0
0.048283
0.218739
8,165
208
133
39.254808
0.665151
0.367789
0
0.783019
0
0
0.056492
0.009118
0
0
0
0
0
1
0.028302
false
0
0.037736
0
0.066038
0.09434
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7