hexsha string | size int64 | ext string | lang string | max_stars_repo_path string | max_stars_repo_name string | max_stars_repo_head_hexsha string | max_stars_repo_licenses list | max_stars_count int64 | max_stars_repo_stars_event_min_datetime string | max_stars_repo_stars_event_max_datetime string | max_issues_repo_path string | max_issues_repo_name string | max_issues_repo_head_hexsha string | max_issues_repo_licenses list | max_issues_count int64 | max_issues_repo_issues_event_min_datetime string | max_issues_repo_issues_event_max_datetime string | max_forks_repo_path string | max_forks_repo_name string | max_forks_repo_head_hexsha string | max_forks_repo_licenses list | max_forks_count int64 | max_forks_repo_forks_event_min_datetime string | max_forks_repo_forks_event_max_datetime string | content string | avg_line_length float64 | max_line_length int64 | alphanum_fraction float64 | qsc_code_num_words_quality_signal int64 | qsc_code_num_chars_quality_signal float64 | qsc_code_mean_word_length_quality_signal float64 | qsc_code_frac_words_unique_quality_signal float64 | qsc_code_frac_chars_top_2grams_quality_signal float64 | qsc_code_frac_chars_top_3grams_quality_signal float64 | qsc_code_frac_chars_top_4grams_quality_signal float64 | qsc_code_frac_chars_dupe_5grams_quality_signal float64 | qsc_code_frac_chars_dupe_6grams_quality_signal float64 | qsc_code_frac_chars_dupe_7grams_quality_signal float64 | qsc_code_frac_chars_dupe_8grams_quality_signal float64 | qsc_code_frac_chars_dupe_9grams_quality_signal float64 | qsc_code_frac_chars_dupe_10grams_quality_signal float64 | qsc_code_frac_chars_replacement_symbols_quality_signal float64 | qsc_code_frac_chars_digital_quality_signal float64 | qsc_code_frac_chars_whitespace_quality_signal float64 | qsc_code_size_file_byte_quality_signal float64 | qsc_code_num_lines_quality_signal float64 | qsc_code_num_chars_line_max_quality_signal float64 | qsc_code_num_chars_line_mean_quality_signal float64 | qsc_code_frac_chars_alphabet_quality_signal float64 | qsc_code_frac_chars_comments_quality_signal float64 | qsc_code_cate_xml_start_quality_signal float64 | qsc_code_frac_lines_dupe_lines_quality_signal float64 | qsc_code_cate_autogen_quality_signal float64 | qsc_code_frac_lines_long_string_quality_signal float64 | qsc_code_frac_chars_string_length_quality_signal float64 | qsc_code_frac_chars_long_word_length_quality_signal float64 | qsc_code_frac_lines_string_concat_quality_signal float64 | qsc_code_cate_encoded_data_quality_signal float64 | qsc_code_frac_chars_hex_words_quality_signal float64 | qsc_code_frac_lines_prompt_comments_quality_signal float64 | qsc_code_frac_lines_assert_quality_signal float64 | qsc_codepython_cate_ast_quality_signal float64 | qsc_codepython_frac_lines_func_ratio_quality_signal float64 | qsc_codepython_cate_var_zero_quality_signal bool | qsc_codepython_frac_lines_pass_quality_signal float64 | qsc_codepython_frac_lines_import_quality_signal float64 | qsc_codepython_frac_lines_simplefunc_quality_signal float64 | qsc_codepython_score_lines_no_logic_quality_signal float64 | qsc_codepython_frac_lines_print_quality_signal float64 | qsc_code_num_words int64 | qsc_code_num_chars int64 | qsc_code_mean_word_length int64 | qsc_code_frac_words_unique null | qsc_code_frac_chars_top_2grams int64 | qsc_code_frac_chars_top_3grams int64 | qsc_code_frac_chars_top_4grams int64 | qsc_code_frac_chars_dupe_5grams int64 | qsc_code_frac_chars_dupe_6grams int64 | qsc_code_frac_chars_dupe_7grams int64 | qsc_code_frac_chars_dupe_8grams int64 | qsc_code_frac_chars_dupe_9grams int64 | qsc_code_frac_chars_dupe_10grams int64 | qsc_code_frac_chars_replacement_symbols int64 | qsc_code_frac_chars_digital int64 | qsc_code_frac_chars_whitespace int64 | qsc_code_size_file_byte int64 | qsc_code_num_lines int64 | qsc_code_num_chars_line_max int64 | qsc_code_num_chars_line_mean int64 | qsc_code_frac_chars_alphabet int64 | qsc_code_frac_chars_comments int64 | qsc_code_cate_xml_start int64 | qsc_code_frac_lines_dupe_lines int64 | qsc_code_cate_autogen int64 | qsc_code_frac_lines_long_string int64 | qsc_code_frac_chars_string_length int64 | qsc_code_frac_chars_long_word_length int64 | qsc_code_frac_lines_string_concat null | qsc_code_cate_encoded_data int64 | qsc_code_frac_chars_hex_words int64 | qsc_code_frac_lines_prompt_comments int64 | qsc_code_frac_lines_assert int64 | qsc_codepython_cate_ast int64 | qsc_codepython_frac_lines_func_ratio int64 | qsc_codepython_cate_var_zero int64 | qsc_codepython_frac_lines_pass int64 | qsc_codepython_frac_lines_import int64 | qsc_codepython_frac_lines_simplefunc int64 | qsc_codepython_score_lines_no_logic int64 | qsc_codepython_frac_lines_print int64 | effective string | hits int64 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
a8f8c437318e16ccc0b9047af193be205971e26a | 23,605 | py | Python | aitoolbox/torchtrain/train_loop/train_loop_tracking.py | mv1388/AIToolbox | c64ac4810a02d230ce471d86b758e82ea232a7e7 | [
"MIT"
] | null | null | null | aitoolbox/torchtrain/train_loop/train_loop_tracking.py | mv1388/AIToolbox | c64ac4810a02d230ce471d86b758e82ea232a7e7 | [
"MIT"
] | 1 | 2019-07-13T21:12:02.000Z | 2019-07-13T21:12:02.000Z | aitoolbox/torchtrain/train_loop/train_loop_tracking.py | mv1388/AIToolbox | c64ac4810a02d230ce471d86b758e82ea232a7e7 | [
"MIT"
] | null | null | null | import os
import inspect
from aitoolbox.torchtrain.train_loop.train_loop import TrainLoop
from aitoolbox.experiment.result_package.abstract_result_packages import AbstractResultPackage
from aitoolbox.torchtrain.callbacks.model_save import ModelCheckpoint, ModelIterationCheckpoint, ModelTrainEndSave
from aitoolbox.torchtrain.train_loop.components.pred_collate_fns import append_predictions, torch_cat_transf
class TrainLoopCheckpoint(TrainLoop):
def __init__(self, model,
train_loader, validation_loader, test_loader,
optimizer, criterion,
project_name, experiment_name, local_model_result_folder_path,
hyperparams,
cloud_save_mode='s3', bucket_name='model-result', cloud_dir_prefix='', source_dirs=(),
rm_subopt_local_models=False, num_best_checkpoints_kept=2,
iteration_save_freq=0,
collate_batch_pred_fn=append_predictions, pred_transform_fn=torch_cat_transf,
end_auto_eval=True, lazy_experiment_save=False,
gpu_mode='single', cuda_device_idx=None, use_amp=False):
"""TrainLoop with the automatic model check-pointing at the end of each epoch
Args:
model (TTModel or ModelWrap or TTDataParallel): neural network model
train_loader (torch.utils.data.DataLoader): data loader for train data set
validation_loader (torch.utils.data.DataLoader or None): data loader for validation data set
test_loader (torch.utils.data.DataLoader or None): data loader for test data set
optimizer (torch.optim.optimizer.Optimizer or MultiOptimizer): optimizer algorithm.
criterion (torch.nn.modules.loss._Loss or MultiLoss or None): criterion during the training procedure
project_name (str): root name of the project
experiment_name (str): name of the particular experiment
local_model_result_folder_path (str): root local path where project folder will be created
hyperparams (dict): used hyper-parameters. When running the TrainLoop from jupyter notebook in order to
ensure the python experiment file copying to the experiment folder, the user needs to manually
specify the python file path as the value for the `experiment_file_path` key. If running the training
directly from the terminal the path deduction is done automatically.
cloud_save_mode (str or None): Storage destination selector.
For AWS S3: 's3' / 'aws_s3' / 'aws'
For Google Cloud Storage: 'gcs' / 'google_storage' / 'google storage'
Everything else results just in local storage to disk
bucket_name (str): name of the bucket in the cloud storage
cloud_dir_prefix (str): path to the folder inside the bucket where the experiments are going to be saved
source_dirs (list or tuple): paths to the local folders with the source code files used in experiment
rm_subopt_local_models (bool or str): if True, the deciding metric is set to 'loss'. Give string metric name
to set it as a deciding metric for suboptimal model removal. If metric name consists of substring 'loss'
the metric minimization is done otherwise metric maximization is done
num_best_checkpoints_kept (int): number of best performing models which are kept when removing suboptimal
model checkpoints
iteration_save_freq (int): frequency of saving the model checkpoint every specified number of
training iterations
collate_batch_pred_fn (callable): collate function transforming batch predictions as they come out from the
model
pred_transform_fn (callable): function transforming all the produced predictions after all the batches have
been run through the model
end_auto_eval (bool or int): used to optionally disable otherwise automatic end of epoch/training val/test
loss calculations. This is useful when conducting very costly experiments to save on compute time.
Specify either True/False boolean to always run or never run after each epoch or specify an int to
execute only every specified number of epochs.
lazy_experiment_save (bool): when in lazy mode experiment tracking components will create the experiment
folder only after some training results are available (possibly at the end of the first epoch) instead
of at the beginning of training.
gpu_mode (str): GPU training mode selection. TrainLoop supports different GPU training modes by
specifying one of the following:
* ``'single'``: single GPU training
* ``'dp'``: multi-GPU training via DataParallel
* ``'ddp'``: multi-GPU training via DistributedDataParallel
cuda_device_idx (int or None): CUDA device index used when training on multiple GPUs
use_amp (bool or dict): use 16-bit Automatic Mixed Precision (AMP)
To switch to AMP mode either:
* set this parameter to ``True`` to use default AMP ``torch.cuda.amp.GradScaler`` initialization params
* provide custom AMP ``torch.cuda.amp.GradScaler`` initialization parameters as a dict as this parameter
"""
TrainLoop.__init__(self, model, train_loader, validation_loader, test_loader, optimizer, criterion,
collate_batch_pred_fn, pred_transform_fn,
end_auto_eval, lazy_experiment_save,
gpu_mode, cuda_device_idx, use_amp)
self.project_name = project_name
self.experiment_name = experiment_name
self.local_model_result_folder_path = os.path.expanduser(local_model_result_folder_path)
self.hyperparams = hyperparams
self.cloud_save_mode = cloud_save_mode
self.bucket_name = bucket_name
self.cloud_dir_prefix = cloud_dir_prefix
self.source_dirs = source_dirs
self.rm_subopt_local_models = rm_subopt_local_models
self.iteration_save_freq = iteration_save_freq
if 'experiment_file_path' not in self.hyperparams:
self.hyperparams['experiment_file_path'] = inspect.getframeinfo(inspect.currentframe().f_back).filename
if 'source_dirs_paths' not in self.hyperparams:
self.hyperparams['source_dirs_paths'] = source_dirs
if iteration_save_freq == 0:
model_checkpoint_cb = ModelCheckpoint(
self.project_name, self.experiment_name, self.local_model_result_folder_path,
self.hyperparams,
cloud_save_mode=self.cloud_save_mode,
bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix,
rm_subopt_local_models=self.rm_subopt_local_models,
num_best_checkpoints_kept=num_best_checkpoints_kept
)
elif iteration_save_freq > 0:
model_checkpoint_cb = ModelIterationCheckpoint(
iteration_save_freq,
self.project_name, self.experiment_name, self.local_model_result_folder_path,
self.hyperparams,
cloud_save_mode=self.cloud_save_mode,
bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix,
rm_subopt_local_models=self.rm_subopt_local_models,
num_best_checkpoints_kept=num_best_checkpoints_kept
)
else:
raise ValueError('iteration_save_freq can have values only >= 0. '
f'But received value {iteration_save_freq}.')
self.callbacks_handler.register_callbacks([model_checkpoint_cb], cache_callbacks=True)
class TrainLoopEndSave(TrainLoop):
def __init__(self, model,
train_loader, validation_loader, test_loader,
optimizer, criterion,
project_name, experiment_name, local_model_result_folder_path,
hyperparams, val_result_package=None, test_result_package=None,
cloud_save_mode='s3', bucket_name='model-result', cloud_dir_prefix='', source_dirs=(),
collate_batch_pred_fn=append_predictions, pred_transform_fn=torch_cat_transf,
end_auto_eval=True, lazy_experiment_save=False,
gpu_mode='single', cuda_device_idx=None, use_amp=False):
"""TrainLoop with the model performance evaluation and final model saving at the end of the training process
Args:
model (TTModel or ModelWrap or TTDataParallel): neural network model
train_loader (torch.utils.data.DataLoader): data loader for train data set
validation_loader (torch.utils.data.DataLoader or None): data loader for validation data set
test_loader (torch.utils.data.DataLoader or None): data loader for test data set
optimizer (torch.optim.optimizer.Optimizer or MultiOptimizer): optimizer algorithm.
criterion (torch.nn.modules.loss._Loss or MultiLoss or None): criterion during the training procedure
project_name (str): root name of the project
experiment_name (str): name of the particular experiment
local_model_result_folder_path (str): root local path where project folder will be created
hyperparams (dict): used hyper-parameters. When running the TrainLoop from jupyter notebook in order to
ensure the python experiment file copying to the experiment folder, the user needs to manually
specify the python file path as the value for the `experiment_file_path` key. If running the training
directly from the terminal the path deduction is done automatically.
val_result_package (AbstractResultPackage or None): result package evaluated on validation data at the end
of the training
test_result_package (AbstractResultPackage or None): result package evaluated on test data at the end
of the training
cloud_save_mode (str or None): Storage destination selector.
For AWS S3: 's3' / 'aws_s3' / 'aws'
For Google Cloud Storage: 'gcs' / 'google_storage' / 'google storage'
Everything else results just in local storage to disk
bucket_name (str): name of the bucket in the cloud storage
cloud_dir_prefix (str): path to the folder inside the bucket where the experiments are going to be saved
source_dirs (list or tuple): paths to the local folders with the source code files used in experiment
collate_batch_pred_fn (callable): collate function transforming batch predictions as they come out from the
model
pred_transform_fn (callable): function transforming all the produced predictions after all the batches have
been run through the model
end_auto_eval (bool or int): used to optionally disable otherwise automatic end of epoch/training val/test
loss calculations. This is useful when conducting very costly experiments to save on compute time.
Specify either True/False boolean to always run or never run after each epoch or specify an int to
execute only every specified number of epochs.
lazy_experiment_save (bool): when in lazy mode experiment tracking components will create the experiment
folder only after some training results are available (possibly at the end of the first epoch) instead
of at the beginning of training.
gpu_mode (str): GPU training mode selection. TrainLoop supports different GPU training modes by
specifying one of the following:
* ``'single'``: single GPU training
* ``'dp'``: multi-GPU training via DataParallel
* ``'ddp'``: multi-GPU training via DistributedDataParallel
cuda_device_idx (int or None): CUDA device index used when training on multiple GPUs
use_amp (bool or dict): use 16-bit Automatic Mixed Precision (AMP)
To switch to AMP mode either:
* set this parameter to ``True`` to use default AMP ``torch.cuda.amp.GradScaler`` initialization params
* provide custom AMP ``torch.cuda.amp.GradScaler`` initialization parameters as a dict as this parameter
"""
TrainLoop.__init__(self, model, train_loader, validation_loader, test_loader, optimizer, criterion,
collate_batch_pred_fn, pred_transform_fn,
end_auto_eval, lazy_experiment_save,
gpu_mode, cuda_device_idx, use_amp)
self.project_name = project_name
self.experiment_name = experiment_name
self.local_model_result_folder_path = os.path.expanduser(local_model_result_folder_path)
self.hyperparams = hyperparams
self.val_result_package = val_result_package
self.test_result_package = test_result_package
self.cloud_save_mode = cloud_save_mode
self.bucket_name = bucket_name
self.cloud_dir_prefix = cloud_dir_prefix
self.source_dirs = source_dirs
if 'experiment_file_path' not in self.hyperparams:
self.hyperparams['experiment_file_path'] = inspect.getframeinfo(inspect.currentframe().f_back).filename
if 'source_dirs_paths' not in self.hyperparams:
self.hyperparams['source_dirs_paths'] = source_dirs
self.check_if_result_packages_possible()
self.callbacks_handler.register_callbacks([
ModelTrainEndSave(self.project_name, self.experiment_name, self.local_model_result_folder_path,
self.hyperparams, self.val_result_package, self.test_result_package,
cloud_save_mode=self.cloud_save_mode,
bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix)
], cache_callbacks=True)
def check_if_result_packages_possible(self):
if self.val_result_package is not None and self.validation_loader is None:
raise ValueError('Given the val_result_package but not supplied the validation_loader. '
'If you want to calculate the val_result_package the validation_loader has to be provided.')
if self.test_result_package is not None and self.test_loader is None:
raise ValueError('Given the test_result_package but not supplied the test_loader. '
'If you want to calculate the test_result_package the test_loader has to be provided.')
if self.val_result_package is None and self.test_result_package is None:
raise ValueError('Both val_result_package and test_result_package are None. '
'At least one of these should be not None but actual result package.')
if self.val_result_package is not None and not isinstance(self.val_result_package, AbstractResultPackage):
raise TypeError(f'val_result_package {self.val_result_package} is not inherited from AbstractResultPackage')
if self.test_result_package is not None and not isinstance(self.test_result_package, AbstractResultPackage):
raise TypeError(f'test_result_package {self.test_result_package} is not inherited from AbstractResultPackage')
class TrainLoopCheckpointEndSave(TrainLoopEndSave):
def __init__(self, model,
train_loader, validation_loader, test_loader,
optimizer, criterion,
project_name, experiment_name, local_model_result_folder_path,
hyperparams, val_result_package=None, test_result_package=None,
cloud_save_mode='s3', bucket_name='model-result', cloud_dir_prefix='', source_dirs=(),
rm_subopt_local_models=False, num_best_checkpoints_kept=2,
iteration_save_freq=0,
collate_batch_pred_fn=append_predictions, pred_transform_fn=torch_cat_transf,
end_auto_eval=True, lazy_experiment_save=False,
gpu_mode='single', cuda_device_idx=None, use_amp=False):
"""TrainLoop both saving model check-pointing at the end of each epoch and model performance reporting
and model saving at the end of the training process
Args:
model (TTModel or ModelWrap or TTDataParallel): neural network model
train_loader (torch.utils.data.DataLoader): data loader for train data set
validation_loader (torch.utils.data.DataLoader or None): data loader for validation data set
test_loader (torch.utils.data.DataLoader or None): data loader for test data set
optimizer (torch.optim.optimizer.Optimizer or MultiOptimizer): optimizer algorithm.
criterion (torch.nn.modules.loss._Loss or MultiLoss or None): criterion during the training procedure
project_name (str): root name of the project
experiment_name (str): name of the particular experiment
local_model_result_folder_path (str): root local path where project folder will be created
hyperparams (dict): used hyper-parameters. When running the TrainLoop from jupyter notebook in order to
ensure the python experiment file copying to the experiment folder, the user needs to manually
specify the python file path as the value for the `experiment_file_path` key. If running the training
directly from the terminal the path deduction is done automatically.
val_result_package (AbstractResultPackage or None): result package evaluated on validation data at the end
of the training
test_result_package (AbstractResultPackage or None): result package evaluated on test data at the end
of the training
cloud_save_mode (str or None): Storage destination selector.
For AWS S3: 's3' / 'aws_s3' / 'aws'
For Google Cloud Storage: 'gcs' / 'google_storage' / 'google storage'
Everything else results just in local storage to disk
bucket_name (str): name of the bucket in the cloud storage
cloud_dir_prefix (str): path to the folder inside the bucket where the experiments are going to be saved
source_dirs (list or tuple): paths to the local folders with the source code files used in experiment
rm_subopt_local_models (bool or str): if True, the deciding metric is set to 'loss'. Give string metric name
to set it as a deciding metric for suboptimal model removal. If metric name consists of substring 'loss'
the metric minimization is done otherwise metric maximization is done
num_best_checkpoints_kept (int): number of best performing models which are kept when removing suboptimal
model checkpoints
iteration_save_freq (int): frequency of saving the model checkpoint every specified number of
training iterations
collate_batch_pred_fn (callable): collate function transforming batch predictions as they come out from the
model
pred_transform_fn (callable): function transforming all the produced predictions after all the batches have
been run through the model
end_auto_eval (bool or int): used to optionally disable otherwise automatic end of epoch/training val/test
loss calculations. This is useful when conducting very costly experiments to save on compute time.
Specify either True/False boolean to always run or never run after each epoch or specify an int to
execute only every specified number of epochs.
lazy_experiment_save (bool): when in lazy mode experiment tracking components will create the experiment
folder only after some training results are available (possibly at the end of the first epoch) instead
of at the beginning of training.
gpu_mode (str): GPU training mode selection. TrainLoop supports different GPU training modes by
specifying one of the following:
* ``'single'``: single GPU training
* ``'dp'``: multi-GPU training via DataParallel
* ``'ddp'``: multi-GPU training via DistributedDataParallel
cuda_device_idx (int or None): CUDA device index used when training on multiple GPUs
use_amp (bool or dict): use 16-bit Automatic Mixed Precision (AMP)
To switch to AMP mode either:
* set this parameter to ``True`` to use default AMP ``torch.cuda.amp.GradScaler`` initialization params
* provide custom AMP ``torch.cuda.amp.GradScaler`` initialization parameters as a dict as this parameter
"""
if 'experiment_file_path' not in hyperparams:
hyperparams['experiment_file_path'] = inspect.getframeinfo(inspect.currentframe().f_back).filename
if 'source_dirs_paths' not in hyperparams:
hyperparams['source_dirs_paths'] = source_dirs
TrainLoopEndSave.__init__(self, model, train_loader, validation_loader, test_loader,
optimizer, criterion,
project_name, experiment_name, os.path.expanduser(local_model_result_folder_path),
hyperparams, val_result_package, test_result_package,
cloud_save_mode, bucket_name, cloud_dir_prefix, source_dirs,
collate_batch_pred_fn, pred_transform_fn,
end_auto_eval, lazy_experiment_save,
gpu_mode, cuda_device_idx, use_amp)
self.rm_subopt_local_models = rm_subopt_local_models
self.iteration_save_freq = iteration_save_freq
if iteration_save_freq == 0:
model_checkpoint_cb = ModelCheckpoint(
self.project_name, self.experiment_name, self.local_model_result_folder_path,
self.hyperparams,
cloud_save_mode=self.cloud_save_mode,
bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix,
rm_subopt_local_models=self.rm_subopt_local_models,
num_best_checkpoints_kept=num_best_checkpoints_kept
)
elif iteration_save_freq > 0:
model_checkpoint_cb = ModelIterationCheckpoint(
iteration_save_freq,
self.project_name, self.experiment_name, self.local_model_result_folder_path,
self.hyperparams,
cloud_save_mode=self.cloud_save_mode,
bucket_name=bucket_name, cloud_dir_prefix=cloud_dir_prefix,
rm_subopt_local_models=self.rm_subopt_local_models,
num_best_checkpoints_kept=num_best_checkpoints_kept
)
else:
raise ValueError('iteration_save_freq can have values only >= 0. '
f'But received value {iteration_save_freq}.')
self.callbacks_handler.register_callbacks([model_checkpoint_cb], cache_callbacks=True)
| 67.83046 | 122 | 0.679178 | 2,933 | 23,605 | 5.237982 | 0.105694 | 0.033848 | 0.01777 | 0.022912 | 0.948448 | 0.937577 | 0.914079 | 0.890516 | 0.889475 | 0.86845 | 0 | 0.001627 | 0.270832 | 23,605 | 347 | 123 | 68.025937 | 0.890948 | 0.505994 | 0 | 0.719745 | 0 | 0 | 0.10234 | 0.01314 | 0 | 0 | 0 | 0 | 0 | 1 | 0.025478 | false | 0 | 0.038217 | 0 | 0.082803 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
d12137748bd2c1bf0580f4cfbbb752d0099513d3 | 853,073 | py | Python | rmgpy/test_data/testing_database/thermo/groups/group.py | pw0908/RMG-Py | 3846fcce701f2a5fd12dbfa429687e9fcd647298 | [
"MIT"
] | 1 | 2022-01-24T05:08:32.000Z | 2022-01-24T05:08:32.000Z | rmgpy/test_data/testing_database/thermo/groups/group.py | speth/RMG-Py | 1d2c2b684580396e984459d9347628a5ceb80e2e | [
"MIT"
] | 72 | 2016-06-06T18:18:49.000Z | 2019-11-17T03:21:10.000Z | rmgpy/test_data/testing_database/thermo/groups/group.py | speth/RMG-Py | 1d2c2b684580396e984459d9347628a5ceb80e2e | [
"MIT"
] | 3 | 2017-09-22T15:47:37.000Z | 2021-12-30T23:51:47.000Z | #!/usr/bin/env python
# encoding: utf-8
name = "Functional Group Additivity Values"
shortDesc = u""
longDesc = u"""
"""
entry(
index = -1,
label = "R",
group =
"""
1 * R u0
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1,
label = "C",
group =
"""
1 * C u0
""",
thermo = u'Cs-CsCsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 2,
label = "Cbf",
group =
"""
1 * Cbf u0
""",
thermo = u'Cbf-CbCbCbf',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 3,
label = "Cbf-CbCbCbf",
group =
"""
1 * Cbf u0 {2,B} {3,B} {4,B}
2 Cb u0 {1,B}
3 Cb u0 {1,B}
4 Cbf u0 {1,B}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.01,3.68,4.2,4.61,5.2,5.7,6.2],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (4.8,'kcal/mol','+|-',0.17),
S298 = (-5,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cbf-CbfCbCb STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 4,
label = "Cbf-CbCbfCbf",
group =
"""
1 * Cbf u0 {2,B} {3,B} {4,B}
2 Cb u0 {1,B}
3 Cbf u0 {1,B}
4 Cbf u0 {1,B}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.01,3.68,4.2,4.61,5.2,5.7,6.2],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (3.7,'kcal/mol','+|-',0.3),
S298 = (-5,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cbf-CbfCbfCb STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 5,
label = "Cbf-CbfCbfCbf",
group =
"""
1 * Cbf u0 p0 c0 {3,B} {6,B} {2,B}
2 Cbf u0 p0 c0 {4,B} {5,B} {1,B}
3 Cbf u0 p0 c0 {8,B} {9,B} {1,B}
4 Cbf u0 p0 c0 {10,B} {11,B} {2,B}
5 Cbf u0 p0 c0 {13,B} {14,B} {2,B}
6 Cbf u0 p0 c0 {15,B} {16,B} {1,B}
7 C u0 p0 c0 {8,B} {16,B}
8 C u0 p0 c0 {7,B} {3,B}
9 C u0 p0 c0 {3,B} {10,B}
10 C u0 p0 c0 {9,B} {4,B}
11 C u0 p0 c0 {4,B} {12,B}
12 C u0 p0 c0 {11,B} {13,B}
13 C u0 p0 c0 {12,B} {5,B}
14 C u0 p0 c0 {5,B} {15,B}
15 C u0 p0 c0 {14,B} {6,B}
16 C u0 p0 c0 {7,B} {6,B}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2,3.11,3.9,4.42,5,5.3,5.7],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (1.5,'kcal/mol','+|-',0.3),
S298 = (1.8,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cbf-CbfCbfCbf STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
The smallest PAH that can have Cbf-CbfCbfCbf is pyrene. Currently the database is restricted
that any group with more three Cbf atoms must have all benzene rings explicitly written out.
Previously, this node would also match one carbon on Benzo[c]phenanthrene and does not now.
Examples from the original source do not include Benzo[c]phenanthrene.
""",
)
entry(
index = 6,
label = "Cb",
group =
"""
1 * Cb u0
""",
thermo = u'Cb-Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 7,
label = "Cb-H",
group =
"""
1 * Cb u0 {2,S}
2 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.24,4.44,5.46,6.3,7.54,8.41,9.73],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (3.3,'kcal/mol','+|-',0.11),
S298 = (11.53,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cb-H BENSON""",
longDesc =
u"""
""",
)
entry(
index = 8,
label = "Cb-O2s",
group =
"""
1 * Cb u0 {2,S}
2 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.9,5.3,6.2,6.6,6.9,6.9,7.07],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-0.9,'kcal/mol','+|-',0.16),
S298 = (-10.2,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cb-O BENSON Cp1500=3D Cp1000*(Cp1500/Cp1000: Cb/Cd)""",
longDesc =
u"""
""",
)
entry(
index = 1197,
label = "Cb-S2s",
group =
"""
1 * Cb u0 {2,S}
2 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2.46,3.24,3.92,4.49,5.27,5.75,6.3],'cal/(mol*K)'),
H298 = (5.83,'kcal/mol'),
S298 = (-7.94,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 9,
label = "Cb-C",
group =
"""
1 * Cb u0 {2,S}
2 C u0 {1,S}
""",
thermo = u'Cb-Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 10,
label = "Cb-Cs",
group =
"""
1 * Cb u0 {2,S}
2 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2.67,3.14,3.68,4.15,4.96,5.44,5.98],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (5.51,'kcal/mol','+|-',0.13),
S298 = (-7.69,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cb-Cs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 11,
label = "Cb-Cds",
group =
"""
1 * Cb u0 {2,S}
2 [Cd,CO] u0 {1,S}
""",
thermo = u'Cb-(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 12,
label = "Cb-(Cds-O2d)",
group =
"""
1 * Cb u0 {2,S}
2 CO u0 {1,S} {3,D}
3 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.59,3.97,4.38,4.72,5.28,5.61,5.75],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (3.69,'kcal/mol','+|-',0.2),
S298 = (-7.8,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Enthalpy from Cb-CO, entropies and heat capacities assigned value of Cb-Cd""",
longDesc =
u"""
""",
)
entry(
index = 13,
label = "Cb-(Cds-Cd)",
group =
"""
1 * Cb u0 {2,S}
2 Cd u0 {1,S} {3,D}
3 C u0 {2,D}
""",
thermo = u'Cb-(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 14,
label = "Cb-(Cds-Cds)",
group =
"""
1 * Cb u0 {2,S}
2 Cd u0 {1,S} {3,D}
3 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.59,3.97,4.38,4.72,5.28,5.61,5.75],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (5.69,'kcal/mol','+|-',0.2),
S298 = (-7.8,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cb-Cd STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 15,
label = "Cb-(Cds-Cdd)",
group =
"""
1 * Cb u0 {2,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D}
""",
thermo = u'Cb-(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 16,
label = "Cb-(Cds-Cdd-O2d)",
group =
"""
1 * Cb u0 {2,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {4,D}
4 O2d u0 {3,D}
""",
thermo = u'Cb-(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cb-(Cds-Cdd-S2d)",
group =
"""
1 * Cb u0 {2,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {4,D}
4 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 17,
label = "Cb-(Cds-Cdd-Cd)",
group =
"""
1 * Cb u0 {2,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {4,D}
4 C u0 {3,D}
""",
thermo = u'Cb-(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 18,
label = "Cb-Ct",
group =
"""
1 * Cb u0 {2,S}
2 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.59,3.97,4.38,4.72,5.28,5.61,5.75],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (5.69,'kcal/mol','+|-',0.3),
S298 = (-7.8,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cb-Ct STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 1839,
label = "Cb-(CtN3t)",
group =
"""
1 * Cb u0 {2,S}
2 Ct u0 {1,S} {3,T}
3 N3t u0 {2,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([9.8,11.2,12.3,13.1,14.2,14.9,16.65],'cal/(mol*K)'),
H298 = (35.8,'kcal/mol'),
S298 = (20.5,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 19,
label = "Cb-Cb",
group =
"""
1 * Cb u0 {2,S}
2 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.33,4.22,4.89,5.27,5.76,5.95,6.05],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (4.96,'kcal/mol','+|-',0.3),
S298 = (-8.64,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cb-Cb BENSON""",
longDesc =
u"""
""",
)
entry(
index = 1821,
label = "Cb-N3s",
group =
"""
1 * Cb u0 {2,S}
2 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.95,5.21,5.94,6.32,6.53,6.56,6.635],'cal/(mol*K)'),
H298 = (-0.5,'kcal/mol'),
S298 = (-9.69,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1198,
label = "Cb-C=S",
group =
"""
1 * Cb u0 {2,S}
2 CS u0 {1,S} {3,D}
3 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2.46,3.24,3.92,4.49,5.27,5.75,6.3],'cal/(mol*K)'),
H298 = (5.83,'kcal/mol'),
S298 = (-7.94,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 20,
label = "Ct",
group =
"""
1 * Ct u0
""",
thermo = u'Ct-CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1849,
label = "Ct-CtN3s",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 N3s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1853,
label = "Ct-N3tN3s",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 N3t u0 {1,T}
3 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 21,
label = "Ct-CtH",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.28,5.99,6.49,6.87,7.47,7.96,8.85],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (26.93,'kcal/mol','+|-',0.05),
S298 = (24.7,'cal/(mol*K)','+|-',0.07),
),
shortDesc = u"""Ct-H STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 22,
label = "Ct-CtOs",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.64,4.39,4.85,5.63,5.66,5.73,5.73],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (31.4,'kcal/mol','+|-',0.27),
S298 = (4.91,'cal/(mol*K)','+|-',0.09),
),
shortDesc = u"""Ct-O MELIUS / hc#coh !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1852,
label = "Ct-N3tOs",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 N3t u0 {1,T}
3 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1195,
label = "Ct-CtSs",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.29,3.67,4,4.29,4.74,5.05,5.49],'cal/(mol*K)'),
H298 = (27.63,'kcal/mol'),
S298 = (6.32,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1942,
label = "Ct-N3tC",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 N3t u0 {1,T}
3 C u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1850,
label = "Ct-N3tCs",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 N3t u0 {1,T}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1851,
label = "Ct-N3tCd",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 N3t u0 {1,T}
3 Cd u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 23,
label = "Ct-CtC",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 C u0 {1,S}
""",
thermo = u'Ct-CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 24,
label = "Ct-CtCs",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.13,3.48,3.81,4.09,4.6,4.92,6.35],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (27.55,'kcal/mol','+|-',0.27),
S298 = (6.35,'cal/(mol*K)','+|-',0.09),
),
shortDesc = u"""Ct-Cs STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 25,
label = "Ct-CtCds",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 [Cd,CO] u0 {1,S}
""",
thermo = u'Ct-Ct(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 26,
label = "Ct-Ct(Cds-O2d)",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 CO u0 {1,S} {4,D}
4 O2d u0 {3,D}
""",
thermo = u'Ct-CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 27,
label = "Ct-Ct(Cds-Cd)",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Cd u0 {1,S} {4,D}
4 C u0 {3,D}
""",
thermo = u'Ct-Ct(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 28,
label = "Ct-Ct(Cds-Cds)",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Cd u0 {1,S} {4,D}
4 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2.57,3.54,3.5,4.92,5.34,5.5,5.8],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (28.2,'kcal/mol','+|-',0.27),
S298 = (6.43,'cal/(mol*K)','+|-',0.09),
),
shortDesc = u"""Ct-Cd STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 29,
label = "Ct-Ct(Cds-Cdd)",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Cd u0 {1,S} {4,D}
4 Cdd u0 {3,D}
""",
thermo = u'Ct-Ct(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 30,
label = "Ct-Ct(Cds-Cdd-O2d)",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Cd u0 {1,S} {4,D}
4 Cdd u0 {3,D} {5,D}
5 O2d u0 {4,D}
""",
thermo = u'Ct-Ct(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Ct-Ct(Cds-Cdd-S2d)",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Cd u0 {1,S} {4,D}
4 Cdd u0 {3,D} {5,D}
5 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 31,
label = "Ct-Ct(Cds-Cdd-Cd)",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Cd u0 {1,S} {4,D}
4 Cdd u0 {3,D} {5,D}
5 C u0 {4,D}
""",
thermo = u'Ct-Ct(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 32,
label = "Ct-CtCt",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.54,4.06,4.4,4.64,5,5.23,5.57],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (25.6,'kcal/mol','+|-',0.27),
S298 = (5.88,'cal/(mol*K)','+|-',0.09),
),
shortDesc = u"""Ct-Ct STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 1840,
label = "Ct-Ct(CtN3t)",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Ct u0 {1,S} {4,T}
4 N3t u0 {3,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.3,11.3,12.1,12.7,13.6,14.3,15.3],'cal/(mol*K)'),
H298 = (63.8,'kcal/mol'),
S298 = (35.4,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 33,
label = "Ct-CtCb",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2.57,3.54,4.5,4.92,5.34,5.5,5.8],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (24.67,'kcal/mol','+|-',0.27),
S298 = (6.43,'cal/(mol*K)','+|-',0.09),
),
shortDesc = u"""Ct-Cb STEIN and FAHR; J. PHYS. CHEM. 1985, 89, 17, 3714""",
longDesc =
u"""
""",
)
entry(
index = 1196,
label = "Ct-CtC=S",
group =
"""
1 * Ct u0 {2,T} {3,S}
2 Ct u0 {1,T}
3 CS u0 {1,S} {4,D}
4 S2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.29,3.67,4,4.29,4.74,5.05,5.49],'cal/(mol*K)'),
H298 = (27.63,'kcal/mol'),
S298 = (6.32,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 34,
label = "Cdd",
group =
"""
1 * Cdd u0
""",
thermo = u'Cdd-CdsCds',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1854,
label = "Cdd-N3dCd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 N3d u0 {1,D}
3 Cd u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.5,5.1,5.6,6,6.5,6.9,7.4],'cal/(mol*K)','+|-',[1.1,1.1,1.1,1.1,1.1,1.1,1.1]),
H298 = (25.9,'kcal/mol','+|-',1.5),
S298 = (19.7,'cal/(mol*K)','+|-',1.4),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 35,
label = "Cdd-OdOd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 O2d u0 {1,D}
3 O2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.91,9.86,10.65,11.31,12.32,12.99,13.93],'cal/(mol*K)'),
H298 = (-94.05,'kcal/mol','+|-',0.03),
S298 = (52.46,'cal/(mol*K)','+|-',0.002),
),
shortDesc = u"""CHEMKIN DATABASE: S(group) = S(CO2) + Rln(2)""",
longDesc =
u"""
""",
)
entry(
index = 1466,
label = "Cdd-OdSd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 O2d u0 {1,D}
3 S2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([9.81,10.8,11.6,12.21,13.03,13.51,14.12],'cal/(mol*K)'),
H298 = (-35.96,'kcal/mol'),
S298 = (55.34,'cal/(mol*K)'),
),
shortDesc = u"""CAC calc 1D-HR""",
longDesc =
u"""
""",
)
entry(
index = 1199,
label = "Cdd-SdSd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 S2d u0 {1,D}
3 S2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.91,11.83,12.49,12.98,13.63,14.01,14.47],'cal/(mol*K)'),
H298 = (24.5,'kcal/mol'),
S298 = (58.24,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2009""",
longDesc =
u"""
""",
)
entry(
index = 36,
label = "Cdd-CdOd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 C u0 {1,D}
3 O2d u0 {1,D}
""",
thermo = u'Cdd-CdsOd',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 37,
label = "Cdd-CdsOd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cd u0 {1,D}
3 O2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""O=C*=C< currently treat the adjacent C as Ck""",
longDesc =
u"""
""",
)
entry(
index = 38,
label = "Cdd-CddOd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D}
3 O2d u0 {1,D}
""",
thermo = u'Cdd-(Cdd-Cd)O2d',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 40,
label = "Cdd-(Cdd-O2d)O2d",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 O2d u0 {1,D}
4 O2d u0 {2,D}
""",
thermo = u'Cdd-CdsOd',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 39,
label = "Cdd-(Cdd-Cd)O2d",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 O2d u0 {1,D}
4 C u0 {2,D}
""",
thermo = u'Cdd-CdsOd',
shortDesc = u"""O=C*=C= currently not defined. Assigned same value as Ca""",
longDesc =
u"""
""",
)
entry(
index = 1200,
label = "Cdd-CdSd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 C u0 {1,D}
3 S2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.88,8.48,8.8,8.99,9.23,9.37,9.58],'cal/(mol*K)'),
H298 = (40.33,'kcal/mol'),
S298 = (34.24,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cdd-CdsSd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cd u0 {1,D}
3 S2d u0 {1,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cdd-CddSd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D}
3 S2d u0 {1,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cdd-(Cdd-S2d)S2d",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 S2d u0 {1,D}
4 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cdd-(Cdd-Cd)S2d",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 S2d u0 {1,D}
4 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 41,
label = "Cdd-CdCd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 C u0 {1,D}
3 C u0 {1,D}
""",
thermo = u'Cdd-CdsCds',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 42,
label = "Cdd-CddCdd",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D}
3 Cdd u0 {1,D}
""",
thermo = u'Cdd-(Cdd-Cd)(Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 43,
label = "Cdd-(Cdd-O2d)(Cdd-O2d)",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 Cdd u0 {1,D} {5,D}
4 O2d u0 {2,D}
5 O2d u0 {3,D}
""",
thermo = u'Cdd-CdsCds',
shortDesc = u"""O=C=C*=C=O, currently not defined. Assigned same value as Ca""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cdd-(Cdd-S2d)(Cdd-S2d)",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 Cdd u0 {1,D} {5,D}
4 S2d u0 {2,D}
5 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 44,
label = "Cdd-(Cdd-O2d)(Cdd-Cd)",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 Cdd u0 {1,D} {5,D}
4 O2d u0 {2,D}
5 C u0 {3,D}
""",
thermo = u'Cdd-(Cdd-O2d)Cds',
shortDesc = u"""O=C=C*=C=C, currently not defined. Assigned same value as Ca""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cdd-(Cdd-S2d)(Cdd-Cd)",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 Cdd u0 {1,D} {5,D}
4 S2d u0 {2,D}
5 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 45,
label = "Cdd-(Cdd-Cd)(Cdd-Cd)",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 Cdd u0 {1,D} {5,D}
4 C u0 {2,D}
5 C u0 {3,D}
""",
thermo = u'Cdd-CdsCds',
shortDesc = u"""C=C=C*=C=C, currently not defined. Assigned same value as Ca""",
longDesc =
u"""
""",
)
entry(
index = 46,
label = "Cdd-CddCds",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D}
3 Cd u0 {1,D}
""",
thermo = u'Cdd-(Cdd-Cd)(Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 47,
label = "Cdd-(Cdd-O2d)Cds",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 Cd u0 {1,D}
4 O2d u0 {2,D}
""",
thermo = u'Cdd-CdsCds',
shortDesc = u"""O=C=C*=C<, currently not defined. Assigned same value as Ca """,
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cdd-(Cdd-S2d)Cds",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 Cd u0 {1,D}
4 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 48,
label = "Cdd-(Cdd-Cd)Cds",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cdd u0 {1,D} {4,D}
3 Cd u0 {1,D}
4 C u0 {2,D}
""",
thermo = u'Cdd-CdsCds',
shortDesc = u"""C=C=C*=C<, currently not defined. Assigned same value as Ca """,
longDesc =
u"""
""",
)
entry(
index = 49,
label = "Cdd-CdsCds",
group =
"""
1 * Cdd u0 {2,D} {3,D}
2 Cd u0 {1,D}
3 Cd u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.9,4.4,4.7,5,5.3,5.5,5.7],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (34.2,'kcal/mol','+|-',0.2),
S298 = (6,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Benson's Ca """,
longDesc =
u"""
""",
)
entry(
index = 50,
label = "Cds",
group =
"""
1 * [Cd,CO,CS] u0
""",
thermo = u'Cds-CdsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1823,
label = "Cds-OdN3sH",
group =
"""
1 * CO u0 {2,S} {3,S} {4,D}
2 N3s u0 {1,S}
3 H u0 {1,S}
4 O2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.03,7.87,8.82,9.68,11.16,12.2,14.8],'cal/(mol*K)'),
H298 = (-29.6,'kcal/mol'),
S298 = (34.93,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1824,
label = "Cds-OdN3sCs",
group =
"""
1 * CO u0 {2,S} {3,S} {4,D}
2 N3s u0 {1,S}
3 Cs u0 {1,S}
4 O2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.37,6.17,7.07,7.66,9.62,11.19,15.115],'cal/(mol*K)'),
H298 = (-32.8,'kcal/mol'),
S298 = (16.2,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1855,
label = "Cd-N3dCsCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 N3d u0 {1,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.5,4.2,5,5.6,6.6,7.2,7.9],'cal/(mol*K)','+|-',[0.9,0.9,0.9,0.9,0.9,0.9,0.9]),
H298 = (5.7,'kcal/mol','+|-',1.2),
S298 = (2,'cal/(mol*K)','+|-',1.1),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1856,
label = "Cd-N3dCsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 N3d u0 {1,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.5,6.3,7.2,8,9.3,10.2,11.6],'cal/(mol*K)','+|-',[0.9,0.9,0.9,0.9,0.9,0.9,0.9]),
H298 = (3.3,'kcal/mol','+|-',1.3),
S298 = (21.2,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1857,
label = "Cd-N3dHH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 N3d u0 {1,D}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.2,7.4,8.7,9.8,11.5,12.9,15],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (4.4,'kcal/mol','+|-',1.4),
S298 = (40.8,'cal/(mol*K)','+|-',1.3),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 51,
label = "Cds-OdHH",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.47,9.38,10.46,11.52,13.37,14.81,14.81],'cal/(mol*K)','+|-',[0.06,0.06,0.06,0.06,0.06,0.06,0.06]),
H298 = (-25.95,'kcal/mol','+|-',0.11),
S298 = (53.68,'cal/(mol*K)','+|-',0.06),
),
shortDesc = u"""CO-HH BENSON !!!WARNING! Cp1500 value taken as Cp1000, S(group) = S(CH2O) + Rln(2)""",
longDesc =
u"""
""",
)
entry(
index = 52,
label = "Cds-OdOsH",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([25.88,34.56,42.08,48.16,56.57,61.38,65.84],'J/(mol*K)','+|-',[4.01,4.01,4.01,4.01,4.01,4.01,4.01]),
H298 = (-211.8,'kJ/mol','+|-',3.42),
S298 = (124.04,'J/(mol*K)','+|-',4.68),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1454,
label = "CO-SsH",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.51,6.16,6.7,7.17,8.06,8.79,9.83],'cal/(mol*K)'),
H298 = (-9.84,'kcal/mol'),
S298 = (29.36,'cal/(mol*K)'),
),
shortDesc = u"""CAC 1d-HR calc""",
longDesc =
u"""
""",
)
entry(
index = 53,
label = "Cds-OdOsOs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([26.17,39.3,48.25,53.88,58.97,59.63,56.09],'J/(mol*K)','+|-',[6,6,6,6,6,6,6]),
H298 = (-281.4,'kJ/mol','+|-',5.11),
S298 = (22.66,'J/(mol*K)','+|-',7),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1455,
label = "CO-CsSs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 S2s u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.37,5.04,5.51,5.83,6.29,6.48,6.38],'cal/(mol*K)'),
H298 = (-14.02,'kcal/mol'),
S298 = (8.55,'cal/(mol*K)'),
),
shortDesc = u"""CAC 1d-HR calc""",
longDesc =
u"""
""",
)
entry(
index = 1456,
label = "CO-OsSs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 O2s u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.94,5.63,6.35,6.99,7.59,7.76,8.18],'cal/(mol*K)'),
H298 = (-11.53,'kcal/mol'),
S298 = (9.61,'cal/(mol*K)'),
),
shortDesc = u"""CAC CBS-QB3 1Dhr calc""",
longDesc =
u"""
""",
)
entry(
index = 54,
label = "Cds-OdCH",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 C u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-OdCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 55,
label = "Cds-OdCsH",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([26.24,31.22,35.94,40.13,46.74,51.39,57.73],'J/(mol*K)','+|-',[2.08,2.08,2.08,2.08,2.08,2.08,2.08]),
H298 = (-123.4,'kJ/mol','+|-',1.77),
S298 = (145.46,'J/(mol*K)','+|-',2.42),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 56,
label = "Cds-OdCdsH",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 57,
label = "Cds-O2d(Cds-O2d)H",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 CO u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([29.76,34.63,39.25,43.32,49.57,53.77,59.32],'J/(mol*K)','+|-',[1.7,1.7,1.7,1.7,1.7,1.7,1.7]),
H298 = (-104.8,'kJ/mol','+|-',1.45),
S298 = (140.49,'J/(mol*K)','+|-',1.98),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 58,
label = "Cds-O2d(Cds-Cd)H",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 H u0 {1,S}
5 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([27.31,34,39.42,43.77,50.16,54.55,60.77],'J/(mol*K)','+|-',[4.9,4.9,4.9,4.9,4.9,4.9,4.9]),
H298 = (-128.3,'kJ/mol','+|-',5.9),
S298 = (129.26,'J/(mol*K)','+|-',5.71),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 59,
label = "Cds-O2d(Cds-Cds)H",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 H u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.45,8.77,9.82,10.7,12.14,12.9,12.9],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (-30.9,'kcal/mol','+|-',0.3),
S298 = (33.4,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""CO-CdH Hf BOZZELLI S,Cp =3D CO/C/H-del(cd syst) !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 60,
label = "Cds-O2d(Cds-Cdd)H",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 H u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 61,
label = "Cds-O2d(Cds-Cdd-O2d)H",
group =
"""
1 * CO u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 O2d u0 {1,D}
5 H u0 {1,S}
6 O2d u0 {3,D}
""",
thermo = u'Cds-O2d(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 62,
label = "Cds-O2d(Cds-Cdd-Cd)H",
group =
"""
1 * CO u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 O2d u0 {1,D}
5 H u0 {1,S}
6 C u0 {3,D}
""",
thermo = u'Cds-O2d(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 63,
label = "Cds-OdCtH",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 64,
label = "Cds-OdCbH",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 65,
label = "Cds-OdCOs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 C u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-OdCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 66,
label = "Cds-OdCsOs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([20.67,28.39,34.6,39.48,46.23,50.09,52.68],'J/(mol*K)','+|-',[2.85,2.85,2.85,2.85,2.85,2.85,2.85]),
H298 = (-222,'kJ/mol','+|-',2.43),
S298 = (43.52,'J/(mol*K)','+|-',3.33),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 67,
label = "Cds-OdCdsOs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 68,
label = "Cds-O2d(Cds-O2d)O2s",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 CO u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([27.18,34.34,39.85,44.13,49.81,52.4,52.33],'J/(mol*K)','+|-',[3.36,3.36,3.36,3.36,3.36,3.36,3.36]),
H298 = (-196.2,'kJ/mol','+|-',2.86),
S298 = (39.37,'J/(mol*K)','+|-',3.92),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 69,
label = "Cds-O2d(Cds-Cd)O2s",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 O2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([28.33,37.84,44.54,49.34,55.45,58.73,60.61],'J/(mol*K)','+|-',[7.46,7.46,7.46,7.46,7.46,7.46,7.46]),
H298 = (-218.6,'kJ/mol','+|-',6.36),
S298 = (33.44,'J/(mol*K)','+|-',8.7),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 70,
label = "Cds-O2d(Cds-Cds)O2s",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 O2s u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.97,6.7,7.4,8.02,8.87,9.36,9.36],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (-32.1,'kcal/mol','+|-',0.3),
S298 = (14.78,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""CO-OCd RPS + S Coreected !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 71,
label = "Cds-O2d(Cds-Cdd)O2s",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 O2s u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 72,
label = "Cds-O2d(Cds-Cdd-O2d)O2s",
group =
"""
1 * CO u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 O2d u0 {1,D}
5 O2s u0 {1,S}
6 O2d u0 {3,D}
""",
thermo = u'Cds-O2d(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 73,
label = "Cds-O2d(Cds-Cdd-Cd)O2s",
group =
"""
1 * CO u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 O2d u0 {1,D}
5 O2s u0 {1,S}
6 C u0 {3,D}
""",
thermo = u'Cds-O2d(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 74,
label = "Cds-OdCtOs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 75,
label = "Cds-OdCbOs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.97,6.7,7.4,8.02,8.87,9.36,9.36],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (-36.6,'kcal/mol','+|-',0.3),
S298 = (14.78,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""CO-OCb RPS + S Coreected !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 76,
label = "Cds-OdCC",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 C u0 {1,S}
4 C u0 {1,S}
""",
thermo = u'Cds-OdCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 77,
label = "Cds-OdCsCs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([23.82,27.7,31.22,34.19,38.37,40.85,43.26],'J/(mol*K)','+|-',[2.08,2.08,2.08,2.08,2.08,2.08,2.08]),
H298 = (-132.2,'kJ/mol','+|-',1.77),
S298 = (61.78,'J/(mol*K)','+|-',2.42),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 78,
label = "Cds-OdCdsCs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 79,
label = "Cds-O2d(Cds-O2d)Cs",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 CO u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([26.77,30.83,34.36,37.27,41.27,43.45,45.25],'J/(mol*K)','+|-',[1.28,1.28,1.28,1.28,1.28,1.28,1.28]),
H298 = (-122,'kJ/mol','+|-',1.09),
S298 = (57.8,'J/(mol*K)','+|-',1.5),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 80,
label = "Cds-O2d(Cds-Cd)Cs",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 Cs u0 {1,S}
5 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([25.26,30.66,34.68,37.69,41.62,43.93,46.69],'J/(mol*K)','+|-',[4.9,4.9,4.9,4.9,4.9,4.9,4.9]),
H298 = (-130.4,'kJ/mol','+|-',4.17),
S298 = (47.38,'J/(mol*K)','+|-',5.71),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 81,
label = "Cds-O2d(Cds-Cds)Cs",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 Cs u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.46,6.32,7.17,7.88,9,9.77,9.77],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (-30.9,'kcal/mol','+|-',0.3),
S298 = (14.6,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""CO-CdCs Hf BENSON =3D CO/Cb/C S,Cp !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 82,
label = "Cds-O2d(Cds-Cdd)Cs",
group =
"""
1 * CO u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 O2d u0 {1,D}
4 Cs u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 83,
label = "Cds-O2d(Cds-Cdd-O2d)Cs",
group =
"""
1 * CO u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 O2d u0 {1,D}
5 Cs u0 {1,S}
6 O2d u0 {3,D}
""",
thermo = u'Cds-O2d(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 84,
label = "Cds-O2d(Cds-Cdd-Cd)Cs",
group =
"""
1 * CO u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 O2d u0 {1,D}
5 Cs u0 {1,S}
6 C u0 {3,D}
""",
thermo = u'Cds-O2d(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 85,
label = "Cds-OdCdsCds",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 86,
label = "Cds-O2d(Cds-O2d)(Cds-O2d)",
group =
"""
1 * CO u0 {2,S} {3,S} {4,D}
2 CO u0 {1,S} {5,D}
3 CO u0 {1,S} {6,D}
4 O2d u0 {1,D}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([31.75,33.35,34.1,34.51,35.19,36.06,38.14],'J/(mol*K)','+|-',[2.41,2.41,2.41,2.41,2.41,2.41,2.41]),
H298 = (-89.3,'kJ/mol','+|-',2.05),
S298 = (64.51,'J/(mol*K)','+|-',2.81),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 87,
label = "Cds-O2d(Cds-Cd)(Cds-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 CO u0 {1,S} {6,D}
5 C u0 {3,D}
6 O2d u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 88,
label = "Cds-O2d(Cds-Cds)(Cds-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 CO u0 {1,S} {6,D}
5 Cd u0 {3,D}
6 O2d u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 89,
label = "Cds-O2d(Cds-Cdd)(Cds-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 CO u0 {1,S} {6,D}
5 Cdd u0 {3,D}
6 O2d u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-Cd)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 90,
label = "Cds-O2d(Cds-Cdd-O2d)(Cds-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 CO u0 {1,S} {7,D}
5 Cdd u0 {3,D} {6,D}
6 O2d u0 {5,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 91,
label = "Cds-O2d(Cds-Cdd-Cd)(Cds-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 CO u0 {1,S} {7,D}
5 Cdd u0 {3,D} {6,D}
6 C u0 {5,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 92,
label = "Cds-O2d(Cds-Cd)(Cds-Cd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 C u0 {3,D}
6 C u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 93,
label = "Cds-O2d(Cds-Cds)(Cds-Cds)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {3,D}
6 Cd u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.46,6.32,7.17,7.88,9,9.77,9.77],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (-30.9,'kcal/mol','+|-',0.3),
S298 = (14.6,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""CO-CdCd Estimate BOZZELLI !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 94,
label = "Cds-O2d(Cds-Cdd)(Cds-Cds)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 Cdd u0 {3,D}
6 Cd u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-Cd)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 95,
label = "Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 Cdd u0 {3,D} {7,D}
6 Cd u0 {4,D}
7 O2d u0 {5,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 96,
label = "Cds-O2d(Cds-Cdd-Cd)(Cds-Cds)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 Cdd u0 {3,D} {7,D}
6 Cd u0 {4,D}
7 C u0 {5,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 97,
label = "Cds-O2d(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 Cdd u0 {3,D}
6 Cdd u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 98,
label = "Cds-O2d(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 Cdd u0 {3,D} {7,D}
6 Cdd u0 {4,D} {8,D}
7 O2d u0 {5,D}
8 O2d u0 {6,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 99,
label = "Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 Cdd u0 {3,D} {7,D}
6 Cdd u0 {4,D} {8,D}
7 C u0 {5,D}
8 O2d u0 {6,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 100,
label = "Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,S} {6,D}
5 Cdd u0 {3,D} {7,D}
6 Cdd u0 {4,D} {8,D}
7 C u0 {5,D}
8 C u0 {6,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 101,
label = "Cds-OdCtCs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 102,
label = "Cds-OdCtCds",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-OdCt(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 103,
label = "Cds-OdCt(Cds-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 CO u0 {1,S} {5,D}
5 O2d u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 104,
label = "Cds-OdCt(Cds-Cd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 C u0 {4,D}
""",
thermo = u'Cds-OdCt(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 105,
label = "Cds-OdCt(Cds-Cds)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 Cd u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 106,
label = "Cds-OdCt(Cds-Cdd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 Cdd u0 {4,D}
""",
thermo = u'Cds-OdCt(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 107,
label = "Cds-OdCt(Cds-Cdd-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 Cdd u0 {4,D} {6,D}
6 O2d u0 {5,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 108,
label = "Cds-OdCt(Cds-Cdd-Cd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 Cdd u0 {4,D} {6,D}
6 C u0 {5,D}
""",
thermo = u'Cds-OdCt(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 109,
label = "Cds-OdCtCt",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 110,
label = "Cds-OdCbCs",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 111,
label = "Cds-OdCbCds",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-OdCb(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 112,
label = "Cds-OdCb(Cds-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 CO u0 {1,S} {5,D}
5 O2d u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 113,
label = "Cds-OdCb(Cds-Cd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 C u0 {4,D}
""",
thermo = u'Cds-OdCb(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 114,
label = "Cds-OdCb(Cds-Cds)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 Cd u0 {4,D}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 115,
label = "Cds-OdCb(Cds-Cdd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 Cdd u0 {4,D}
""",
thermo = u'Cds-OdCb(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 116,
label = "Cds-OdCb(Cds-Cdd-O2d)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 Cdd u0 {4,D} {6,D}
6 O2d u0 {5,D}
""",
thermo = u'Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 117,
label = "Cds-OdCb(Cds-Cdd-Cd)",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 Cd u0 {1,S} {5,D}
5 Cdd u0 {4,D} {6,D}
6 C u0 {5,D}
""",
thermo = u'Cds-OdCb(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 118,
label = "Cds-OdCbCt",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
""",
thermo = u'Cds-OdCt(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 119,
label = "Cds-OdCbCb",
group =
"""
1 * CO u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
""",
thermo = u'Cds-O2d(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 120,
label = "Cds-CdHH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-CdsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 121,
label = "Cds-CdsHH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.1,6.36,7.51,8.5,10.07,11.27,13.19],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (6.26,'kcal/mol','+|-',0.19),
S298 = (27.61,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-HH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 122,
label = "Cds-CddHH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 123,
label = "Cds-(Cdd-O2d)HH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.08,13.91,15.4,16.64,18.61,20.1,22.47],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (-11.34,'kcal/mol','+|-',0.19),
S298 = (57.47,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{CCO/H2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)HH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 124,
label = "Cds-(Cdd-Cd)HH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 125,
label = "Cds-CdOsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([18.08,21.17,24.43,27.41,32.22,35.73,40.97],'J/(mol*K)'),
H298 = (36.4,'kJ/mol'),
S298 = (33.51,'J/(mol*K)'),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 126,
label = "Cds-CdsOsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.75,6.46,7.64,8.35,9.1,9.56,10.46],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (2.03,'kcal/mol','+|-',0.19),
S298 = (6.2,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-OH BOZZELLI Hf vin-oh RADOM + C/Cd/H, S&Cp LAY""",
longDesc =
u"""
""",
)
entry(
index = 127,
label = "Cds-CddOsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 128,
label = "Cds-(Cdd-O2d)OsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.29,13.67,15.1,16.1,17.36,18.25,19.75],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (2.11,'kcal/mol','+|-',0.19),
S298 = (38.17,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{CCO/O/H} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 129,
label = "Cds-(Cdd-Cd)OsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdSsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1180,
label = "Cds-CdsSsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.41,5.2,5.98,6.68,7.8,8.62,9.84],'cal/(mol*K)'),
H298 = (8.87,'kcal/mol'),
S298 = (7.87,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CddSsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)SsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)SsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 130,
label = "Cds-CdOsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.34,11.93,14.86,17.95,22.31,24.6,26.92],'J/(mol*K)','+|-',[7.4,7.4,7.4,7.4,7.4,7.4,7.4]),
H298 = (28.3,'kJ/mol','+|-',6.3),
S298 = (-42.69,'J/(mol*K)','+|-',8.63),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 131,
label = "Cds-CdsOsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-CdsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 132,
label = "Cds-CddOsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 133,
label = "Cds-(Cdd-O2d)OsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.56,15.58,17.69,18.67,18.78,18.4,18.01],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (2.403,'kcal/mol','+|-',0.19),
S298 = (13.42,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{CCO/O2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 134,
label = "Cds-(Cdd-Cd)OsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdSsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsSsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CddSsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)SsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)SsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 135,
label = "Cds-CdCH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 C u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-CdsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 136,
label = "Cds-CdsCsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.16,5.03,5.81,6.5,7.65,8.45,9.62],'cal/(mol*K)','+|-',[0.06,0.06,0.06,0.06,0.06,0.06,0.06]),
H298 = (8.59,'kcal/mol','+|-',0.17),
S298 = (7.97,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-CsH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 137,
label = "Cds-CdsCdsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-Cds(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 139,
label = "Cds-Cds(Cds-Cd)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 H u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 140,
label = "Cds-Cds(Cds-Cds)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 H u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.46,5.79,6.75,7.42,8.35,8.99,9.98],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (6.78,'kcal/mol','+|-',0.2),
S298 = (6.38,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-CdH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 141,
label = "Cds-Cds(Cds-Cdd)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 H u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd-S2d)H",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 H u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 143,
label = "Cds-Cds(Cds-Cdd-Cd)H",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 H u0 {1,S}
6 C u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 144,
label = "Cds-CdsCtH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.46,5.79,6.75,7.42,8.35,8.99,9.98],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (6.78,'kcal/mol','+|-',0.2),
S298 = (6.38,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-CtH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 1836,
label = "Cds-CdsH(CtN3t)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Ct u0 {1,S} {5,T}
3 Cd u0 {1,D}
4 H u0 {1,S}
5 N3t u0 {2,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.3,12,13.4,14.6,16.3,17.5,19.4],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (38.5,'kcal/mol','+|-',1.3),
S298 = (37.6,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 145,
label = "Cds-CdsCbH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.46,5.79,6.75,7.42,8.35,8.99,9.98],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (6.78,'kcal/mol','+|-',0.2),
S298 = (6.38,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-CbH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 146,
label = "Cds-CddCsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 147,
label = "Cds-(Cdd-O2d)CsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([43.83,50.1,55.5,60.05,67.09,72.13,79.55],'J/(mol*K)','+|-',[4,4,4,4,4,4,4]),
H298 = (-17.6,'kJ/mol','+|-',3.41),
S298 = (169.15,'J/(mol*K)','+|-',4.67),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 148,
label = "Cds-(Cdd-Cd)CsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 149,
label = "Cds-CddCdsH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 150,
label = "Cds-(Cdd-O2d)(Cds-O2d)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {6,D}
4 H u0 {1,S}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 151,
label = "Cds-(Cdd-O2d)(Cds-Cd)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 O2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([43.67,52.95,59.65,64.67,71.81,76.72,83.92],'J/(mol*K)','+|-',[5.66,5.66,5.66,5.66,5.66,5.66,5.66]),
H298 = (-36,'kJ/mol','+|-',4.82),
S298 = (152.19,'J/(mol*K)','+|-',6.6),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 152,
label = "Cds-(Cdd-O2d)(Cds-Cds)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 O2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 153,
label = "Cds-(Cdd-O2d)(Cds-Cdd)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 O2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 154,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 H u0 {1,S}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.55,12.41,13.82,14.91,16.51,17.62,19.24],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-4.998,'kcal/mol','+|-',0.2),
S298 = (39.06,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{CCO/H/CCO} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 155,
label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 H u0 {1,S}
6 O2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cd)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 S2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cds)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 S2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 S2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 H u0 {1,S}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 H u0 {1,S}
6 S2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 156,
label = "Cds-(Cdd-Cd)(Cds-O2d)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {6,D}
4 H u0 {1,S}
5 C u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = u'Cd-Cd(CO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 157,
label = "Cds-(Cdd-Cd)(Cds-Cd)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 C u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 158,
label = "Cds-(Cdd-Cd)(Cds-Cds)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 C u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 159,
label = "Cds-(Cdd-Cd)(Cds-Cdd)H",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 C u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 160,
label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 H u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cd-Cd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 H u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 161,
label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)H",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 H u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 162,
label = "Cds-CddCtH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 163,
label = "Cds-(Cdd-O2d)CtH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CtH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 164,
label = "Cds-(Cdd-Cd)CtH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 165,
label = "Cds-CddCbH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 166,
label = "Cds-(Cdd-O2d)CbH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CbH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 167,
label = "Cds-(Cdd-Cd)CbH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)C=SH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {6,D}
4 H u0 {1,S}
5 C u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)C=SH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {6,D}
4 H u0 {1,S}
5 S2d u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1182,
label = "Cds-CdsC=SH",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 H u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.41,5.2,5.98,6.68,7.8,8.62,9.84],'cal/(mol*K)'),
H298 = (8.87,'kcal/mol'),
S298 = (7.87,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1999,
label = "Cd-Cd(CO)H",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CO u0 {1,S} {5,D}
3 H u0 {1,S}
4 Cd u0 {1,D}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([18.08,21.17,24.43,27.41,32.22,35.73,40.97],'J/(mol*K)'),
H298 = (36.4,'kJ/mol'),
S298 = (33.51,'J/(mol*K)'),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 2000,
label = "Cd-Cd(CCO)H",
group =
"""
1 * Cd u0 {2,S} {4,S} {5,D}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 H u0 {1,S}
5 Cd u0 {1,D}
6 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([18.08,21.17,24.43,27.41,32.22,35.73,40.97],'J/(mol*K)'),
H298 = (36.4,'kJ/mol'),
S298 = (33.51,'J/(mol*K)'),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 168,
label = "Cds-CdCO",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 C u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-CdsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 170,
label = "Cds-CdsCdsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-Cds(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 171,
label = "Cds-Cds(Cds-O2d)O2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 CO u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (5.13,'kcal/mol','+|-',0.2),
S298 = (-14.6,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-OCO adj BENSON for RADOM c*coh""",
longDesc =
u"""
""",
)
entry(
index = 172,
label = "Cds-Cds(Cds-Cd)O2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 O2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 173,
label = "Cds-Cds(Cds-Cds)O2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 O2s u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (1.5,'kcal/mol','+|-',0.2),
S298 = (-14.4,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-OCd jwb need calc""",
longDesc =
u"""
""",
)
entry(
index = 174,
label = "Cds-Cds(Cds-Cdd)O2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 O2s u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 175,
label = "Cds-Cds(Cds-Cdd-O2d)O2s",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 O2s u0 {1,S}
6 O2d u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 176,
label = "Cds-Cds(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 O2s u0 {1,S}
6 C u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 177,
label = "Cds-CdsCtOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-Cds(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 178,
label = "Cds-CdsCbOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (1.5,'kcal/mol','+|-',0.2),
S298 = (-14.4,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-OCb jwb need calc""",
longDesc =
u"""
""",
)
entry(
index = 182,
label = "Cds-CddCdsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 183,
label = "Cds-(Cdd-O2d)(Cds-O2d)O2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 184,
label = "Cds-(Cdd-O2d)(Cds-Cd)O2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 185,
label = "Cds-(Cdd-O2d)(Cds-Cds)O2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 186,
label = "Cds-(Cdd-O2d)(Cds-Cdd)O2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 187,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)O2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 O2s u0 {1,S}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.01,12.97,14.17,14.97,15.8,16.26,16.88],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (1.607,'kcal/mol','+|-',0.2),
S298 = (17.73,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{CCO/O/CCO} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 188,
label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 O2s u0 {1,S}
6 O2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 189,
label = "Cds-(Cdd-Cd)(Cds-Cd)O2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 C u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 190,
label = "Cds-(Cdd-Cd)(Cds-Cds)O2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 C u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 191,
label = "Cds-(Cdd-Cd)(Cds-Cdd)O2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 C u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 192,
label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)O2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 O2s u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 193,
label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 O2s u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 194,
label = "Cds-CddCtOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 195,
label = "Cds-(Cdd-O2d)CtOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 196,
label = "Cds-(Cdd-Cd)CtOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 197,
label = "Cds-CddCbOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 198,
label = "Cds-(Cdd-O2d)CbOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 199,
label = "Cds-(Cdd-Cd)CbOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1995,
label = "Cd-CdCsOs",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cs u0 {1,S}
3 O2s u0 {1,S}
4 C u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.79,15.86,19.67,22.91,26.55,27.85,28.45],'J/(mol*K)','+|-',[5.1,5.1,5.1,5.1,5.1,5.1,5.1]),
H298 = (33,'kJ/mol','+|-',4.34),
S298 = (-50.89,'J/(mol*K)','+|-',5.94),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 169,
label = "Cds-CdsCsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.59,4.56,5.04,5.3,5.84,6.07,6.16],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (3.03,'kcal/mol','+|-',0.2),
S298 = (-12.32,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cd-OCs BOZZELLI-RADOM vin-oh and del (ccoh-ccohc)""",
longDesc =
u"""
""",
)
entry(
index = 179,
label = "Cds-CddCsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 180,
label = "Cds-(Cdd-O2d)CsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.91,12.65,13.59,14.22,15,15.48,16.28],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (3.273,'kcal/mol','+|-',0.2),
S298 = (18.58,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{CCO/O/C} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 181,
label = "Cds-(Cdd-Cd)CsOs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdCS",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 C u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1181,
label = "Cds-CdsCsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.23,4.63,4.97,5.29,5.83,6.17,6.53],'cal/(mol*K)'),
H298 = (10.63,'kcal/mol'),
S298 = (-12.76,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsCdsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cd)S2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 S2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cds)S2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 S2s u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd)S2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 S2s u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd-S2d)S2s",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 S2s u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 S2s u0 {1,S}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsCtSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsCbSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CddCsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)CsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CddCdsSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cd)S2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cds)S2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd)S2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)S2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 S2s u0 {1,S}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 S2s u0 {1,S}
6 S2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cd)S2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 C u0 {2,D}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cds)S2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 C u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd)S2s",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 C u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)S2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 S2s u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 S2s u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CddCtSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CtSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)CtSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CddCbSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CbSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)CbSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)C=SSs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsC=SSs",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 200,
label = "Cds-CdCC",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 C u0 {1,S}
4 C u0 {1,S}
""",
thermo = u'Cds-CdsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 201,
label = "Cds-CdsCsCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.1,4.61,4.99,5.26,5.8,6.08,6.36],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (10.34,'kcal/mol','+|-',0.24),
S298 = (-12.7,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CsCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 202,
label = "Cds-CdsCdsCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = u'Cds-Cds(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 204,
label = "Cds-Cds(Cds-Cd)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cs u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 205,
label = "Cds-Cds(Cds-Cds)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cs u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (8.88,'kcal/mol','+|-',0.24),
S298 = (-14.6,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CdCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 206,
label = "Cds-Cds(Cds-Cdd)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cs u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd-S2d)Cs",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 Cs u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 208,
label = "Cds-Cds(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 Cs u0 {1,S}
6 C u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 209,
label = "Cds-CdsCdsCds",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 210,
label = "Cds-Cds(Cds-O2d)(Cds-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CO u0 {1,S} {5,D}
3 CO u0 {1,S} {6,D}
4 Cd u0 {1,D}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = u'Cds-CdsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 211,
label = "Cds-Cds(Cds-O2d)(Cds-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CO u0 {1,S} {6,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,D}
5 C u0 {3,D}
6 O2d u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 212,
label = "Cds-Cds(Cds-O2d)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CO u0 {1,S} {6,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,D}
5 Cd u0 {3,D}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.7,6.13,6.87,7.1,7.2,7.16,7.06],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (11.6,'kcal/mol','+|-',0.24),
S298 = (-16.5,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-COCd from CD/CD2/ jwb est 6/97""",
longDesc =
u"""
AG Vandeputte, added 7 kcal/mol to the following value (see phd M Sabbe)
""",
)
entry(
index = 213,
label = "Cds-Cds(Cds-O2d)(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CO u0 {1,S} {6,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,D}
5 Cdd u0 {3,D}
6 O2d u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-O2d)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 214,
label = "Cds-Cds(Cds-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cd u0 {1,D}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cd-CdCs(CCO)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 215,
label = "Cds-Cds(Cds-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cd u0 {1,D}
6 O2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 216,
label = "Cds-Cds(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,D}
5 C u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 217,
label = "Cds-Cds(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,D}
5 Cd u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([1.9,2.69,3.5,4.28,5.57,6.21,7.37],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (11.6,'kcal/mol','+|-',0.24),
S298 = (-15.67,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CdCd Hf=3D est S,Cp mopac nov99""",
longDesc =
u"""
AG Vandeputte, added 7 kcal/mol to the following value (see phd M Sabbe)
""",
)
entry(
index = 218,
label = "Cds-Cds(Cds-Cds)(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,D}
5 Cd u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 219,
label = "Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cd u0 {1,D}
6 Cd u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cd-CdCs(CCO)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cds)(Cds-Cdd-S2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cd u0 {1,D}
6 Cd u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 220,
label = "Cds-Cds(Cds-Cds)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cd u0 {1,D}
6 Cd u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 221,
label = "Cds-Cds(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,D}
5 Cdd u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 222,
label = "Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {6,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {7,D}
5 Cdd u0 {3,D} {8,D}
6 Cd u0 {1,D}
7 O2d u0 {4,D}
8 O2d u0 {5,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 223,
label = "Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {6,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {7,D}
5 Cdd u0 {3,D} {8,D}
6 Cd u0 {1,D}
7 O2d u0 {4,D}
8 C u0 {5,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {6,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {7,D}
5 Cdd u0 {3,D} {8,D}
6 Cd u0 {1,D}
7 S2d u0 {4,D}
8 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {6,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {7,D}
5 Cdd u0 {3,D} {8,D}
6 Cd u0 {1,D}
7 S2d u0 {4,D}
8 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 224,
label = "Cds-Cds(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {6,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {7,D}
5 Cdd u0 {3,D} {8,D}
6 Cd u0 {1,D}
7 C u0 {4,D}
8 C u0 {5,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 225,
label = "Cds-CdsCtCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.5,3.88,4.88,4.18,4.86,5.4,6.01],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (8.11,'kcal/mol','+|-',0.24),
S298 = (-13.02,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CtCs RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 1858,
label = "Cd-CdCs(CtN3t)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D} {5,S} {6,S}
3 Ct u0 {1,S} {7,T}
4 Cs u0 {1,S}
5 R u0 {2,S}
6 R u0 {2,S}
7 N3t u0 {3,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([9.2,10.6,11.7,12.5,13.8,14.7,15.9],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (40.2,'kcal/mol','+|-',1.3),
S298 = (17.9,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 226,
label = "Cds-CdsCtCds",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-Cds(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 227,
label = "Cds-CdsCt(Cds-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 CO u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Ct u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 228,
label = "Cds-CdsCt(Cds-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Ct u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 229,
label = "Cds-Cds(Cds-Cds)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Ct u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.89,5.26,5.98,6.37,6.67,6.78,6.89],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (7.54,'kcal/mol','+|-',0.24),
S298 = (-14.65,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CtCd RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 230,
label = "Cds-Cds(Cds-Cdd)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Ct u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-Cd)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 231,
label = "Cds-Cds(Cds-Cdd-O2d)Ct",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 Ct u0 {1,S}
6 O2d u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd-S2d)Ct",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 Ct u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 232,
label = "Cds-Cds(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 Ct u0 {1,S}
6 C u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 233,
label = "Cds-CdsCtCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.23,4.59,5.41,5.93,6.48,6.74,7.02],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (8.81,'kcal/mol','+|-',0.24),
S298 = (-13.51,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CtCt RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 1837,
label = "Cds-Cd(CtN3t)(CtN3t)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Ct u0 {1,S} {5,T}
3 Ct u0 {1,S} {6,T}
4 Cd u0 {1,D}
5 N3t u0 {2,T}
6 N3t u0 {3,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (84.1,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 234,
label = "Cds-CdsCbCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.4,5.37,5.93,6.18,6.5,6.62,6.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (8.64,'kcal/mol','+|-',0.24),
S298 = (-14.6,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CbCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 235,
label = "Cds-CdsCbCds",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cb u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-Cds(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 236,
label = "Cds-CdsCb(Cds-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 CO u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cb u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 237,
label = "Cds-Cds(Cds-Cd)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cb u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 238,
label = "Cds-Cds(Cds-Cds)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cb u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.7,6.13,6.87,7.1,7.2,7.16,7.06],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (7.18,'kcal/mol','+|-',0.24),
S298 = (-16.5,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CbCd BOZZELLI =3D Cd/Cs/Cb + (Cd/Cs/Cd - Cd/Cs/Cs)""",
longDesc =
u"""
""",
)
entry(
index = 239,
label = "Cds-Cds(Cds-Cdd)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cb u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-Cd)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 240,
label = "Cds-Cds(Cds-Cdd-O2d)Cb",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 Cb u0 {1,S}
6 O2d u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-Cds(Cds-Cdd-S2d)Cb",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 Cb u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 241,
label = "Cds-Cds(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cd u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cd u0 {1,D}
5 Cb u0 {1,S}
6 C u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 242,
label = "Cds-CdsCbCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2.22,3.14,4.54,4.11,5.06,5.79,6.71],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (6.7,'kcal/mol','+|-',0.24),
S298 = (-17.04,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CbCt Hf=3D est S,Cp mopac nov99""",
longDesc =
u"""
""",
)
entry(
index = 243,
label = "Cds-CdsCbCb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.7,6.13,6.87,7.1,7.2,7.16,7.06],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (8,'kcal/mol','+|-',0.24),
S298 = (-16.5,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cd-CbCb BOZZELLI =3D Cd/Cs/Cb + (Cd/Cs/Cb - Cd/Cs/Cs)""",
longDesc =
u"""
""",
)
entry(
index = 244,
label = "Cds-CddCsCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 245,
label = "Cds-(Cdd-O2d)CsCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([42.55,46.42,50,53.24,58.3,61.71,66.01],'J/(mol*K)','+|-',[4.76,4.76,4.76,4.76,4.76,4.76,4.76]),
H298 = (0.5,'kJ/mol','+|-',4.06),
S298 = (84.72,'J/(mol*K)','+|-',5.55),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CsCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 246,
label = "Cds-(Cdd-Cd)CsCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 247,
label = "Cds-CddCdsCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 248,
label = "Cds-(Cdd-O2d)(Cds-O2d)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 249,
label = "Cds-(Cdd-O2d)(Cds-Cd)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 250,
label = "Cds-(Cdd-O2d)(Cds-Cds)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 251,
label = "Cds-(Cdd-O2d)(Cds-Cdd)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 252,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cs u0 {1,S}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.1,11.24,12.12,12.84,14,14.75,15.72],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-2.07,'kcal/mol','+|-',0.24),
S298 = (19.65,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""{CCO/C/CCO} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 253,
label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cs u0 {1,S}
6 O2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cd)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cds)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cs u0 {1,S}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cs u0 {1,S}
6 S2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 254,
label = "Cds-(Cdd-Cd)(Cds-Cd)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 C u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 255,
label = "Cds-(Cdd-Cd)(Cds-Cds)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 C u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 256,
label = "Cds-(Cdd-Cd)(Cds-Cdd)Cs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 C u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 257,
label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cd-CdCs(CCO)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 258,
label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 259,
label = "Cds-CddCdsCds",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 260,
label = "Cds-(Cdd-O2d)(Cds-O2d)(Cds-O2d)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {6,D}
4 CO u0 {1,S} {7,D}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 261,
label = "Cds-(Cdd-O2d)(Cds-Cd)(Cds-O2d)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {7,D}
5 O2d u0 {2,D}
6 C u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 262,
label = "Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {7,D}
5 O2d u0 {2,D}
6 Cd u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 263,
label = "Cds-(Cdd-O2d)(Cds-Cdd)(Cds-O2d)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {7,D}
5 O2d u0 {2,D}
6 Cdd u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 264,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 CO u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
8 O2d u0 {5,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 265,
label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 CO u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
8 C u0 {5,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 266,
label = "Cds-(Cdd-O2d)(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 O2d u0 {2,D}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 267,
label = "Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 O2d u0 {2,D}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 268,
label = "Cds-(Cdd-O2d)(Cds-Cdd)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 O2d u0 {2,D}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 269,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 O2d u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {5,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 270,
label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 O2d u0 {3,D}
7 Cd u0 {4,D}
8 C u0 {5,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 271,
label = "Cds-(Cdd-O2d)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 O2d u0 {2,D}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 272,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 O2d u0 {4,D}
8 O2d u0 {5,D}
9 O2d u0 {6,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 273,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 O2d u0 {4,D}
8 O2d u0 {5,D}
9 C u0 {6,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 274,
label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 O2d u0 {4,D}
8 C u0 {5,D}
9 C u0 {6,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 275,
label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-O2d)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {6,D}
4 CO u0 {1,S} {7,D}
5 C u0 {2,D}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-Cds(Cds-O2d)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 276,
label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {7,D}
4 Cd u0 {1,S} {6,D}
5 C u0 {2,D}
6 C u0 {4,D}
7 O2d u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 277,
label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {7,D}
4 Cd u0 {1,S} {6,D}
5 C u0 {2,D}
6 Cd u0 {4,D}
7 O2d u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 278,
label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {7,D}
4 Cd u0 {1,S} {6,D}
5 C u0 {2,D}
6 Cdd u0 {4,D}
7 O2d u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 279,
label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 CO u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 C u0 {3,D}
7 O2d u0 {4,D}
8 O2d u0 {5,D}
""",
thermo = u'Cds-Cds(Cds-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 280,
label = "Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 CO u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 C u0 {3,D}
7 O2d u0 {4,D}
8 C u0 {5,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 S2d u0 {2,D}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 S2d u0 {2,D}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 S2d u0 {2,D}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 S2d u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 S2d u0 {3,D}
7 Cd u0 {4,D}
8 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 S2d u0 {2,D}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 S2d u0 {4,D}
8 S2d u0 {5,D}
9 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 S2d u0 {4,D}
8 S2d u0 {5,D}
9 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 S2d u0 {4,D}
8 C u0 {5,D}
9 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 281,
label = "Cds-(Cdd-Cd)(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 C u0 {2,D}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 282,
label = "Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 C u0 {2,D}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 283,
label = "Cds-(Cdd-Cd)(Cds-Cdd)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 C u0 {2,D}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 284,
label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 C u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {5,D}
""",
thermo = u'Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 C u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 285,
label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 C u0 {3,D}
7 Cd u0 {4,D}
8 C u0 {5,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 286,
label = "Cds-(Cdd-Cd)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 C u0 {2,D}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 287,
label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 C u0 {4,D}
8 O2d u0 {5,D}
9 O2d u0 {6,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 288,
label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 C u0 {4,D}
8 O2d u0 {5,D}
9 C u0 {6,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 C u0 {4,D}
8 S2d u0 {5,D}
9 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 C u0 {4,D}
8 S2d u0 {5,D}
9 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 289,
label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {1,D} {7,D}
5 Cdd u0 {2,D} {8,D}
6 Cdd u0 {3,D} {9,D}
7 C u0 {4,D}
8 C u0 {5,D}
9 C u0 {6,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 290,
label = "Cds-CddCtCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 291,
label = "Cds-(Cdd-O2d)CtCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CtCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 292,
label = "Cds-(Cdd-Cd)CtCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 293,
label = "Cds-CddCtCds",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 294,
label = "Cds-(Cdd-O2d)(Cds-O2d)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 295,
label = "Cds-(Cdd-O2d)(Cds-Cd)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 O2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 296,
label = "Cds-(Cdd-O2d)(Cds-Cds)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 O2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 297,
label = "Cds-(Cdd-O2d)(Cds-Cdd)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 O2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 298,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Ct u0 {1,S}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 299,
label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Ct u0 {1,S}
6 O2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cd)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 S2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cds)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 S2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 S2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Ct u0 {1,S}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Ct u0 {1,S}
6 S2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 300,
label = "Cds-(Cdd-Cd)(Cds-Cd)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 C u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 301,
label = "Cds-(Cdd-Cd)(Cds-Cds)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 C u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 302,
label = "Cds-(Cdd-Cd)(Cds-Cdd)Ct",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 C u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 303,
label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-O2d)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 304,
label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 305,
label = "Cds-CddCtCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 306,
label = "Cds-(Cdd-O2d)CtCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CtCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 307,
label = "Cds-(Cdd-Cd)CtCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 308,
label = "Cds-CddCbCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 309,
label = "Cds-(Cdd-O2d)CbCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CbCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 310,
label = "Cds-(Cdd-Cd)CbCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 311,
label = "Cds-CddCbCds",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cb u0 {1,S}
4 [Cd,CO] u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 312,
label = "Cds-(Cdd-O2d)(Cds-O2d)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CO u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 313,
label = "Cds-(Cdd-O2d)(Cds-Cd)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 O2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 314,
label = "Cds-(Cdd-O2d)(Cds-Cds)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 O2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 315,
label = "Cds-(Cdd-O2d)(Cds-Cdd)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 O2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 316,
label = "Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cb u0 {1,S}
6 O2d u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 317,
label = "Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cb u0 {1,S}
6 O2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cd)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 S2d u0 {2,D}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cds)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 S2d u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 S2d u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cb u0 {1,S}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cb u0 {1,S}
6 S2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 318,
label = "Cds-(Cdd-Cd)(Cds-Cd)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 C u0 {2,D}
6 C u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 319,
label = "Cds-(Cdd-Cd)(Cds-Cds)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 C u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = u'Cds-Cds(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 320,
label = "Cds-(Cdd-Cd)(Cds-Cdd)Cb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 C u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 321,
label = "Cds-(Cdd-Cd)(Cds-Cdd-O2d)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cb u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {4,D}
""",
thermo = u'Cds-Cds(Cds-Cdd-O2d)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)(Cds-Cdd-S2d)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cb u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 322,
label = "Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cd u0 {2,S} {3,D} {5,S}
2 Cd u0 {1,S} {4,D}
3 Cdd u0 {1,D} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cb u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
""",
thermo = u'Cds-(Cdd-Cd)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 323,
label = "Cds-CddCbCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 324,
label = "Cds-(Cdd-O2d)CbCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CbCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 325,
label = "Cds-(Cdd-Cd)CbCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 326,
label = "Cds-CddCbCb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
""",
thermo = u'Cds-(Cdd-Cd)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 327,
label = "Cds-(Cdd-O2d)CbCb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = u'Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)CbCb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 328,
label = "Cds-(Cdd-Cd)CbCb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 C u0 {2,D}
""",
thermo = u'Cds-CdsCbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsC=SC=S",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CS u0 {1,S} {5,D}
3 CS u0 {1,S} {6,D}
4 Cd u0 {1,D}
5 S2d u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)C=S(Cds-Cd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {7,D}
4 Cd u0 {1,S} {6,D}
5 C u0 {2,D}
6 C u0 {4,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)C=S(Cds-Cds)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {7,D}
4 Cd u0 {1,S} {6,D}
5 C u0 {2,D}
6 Cd u0 {4,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)C=S(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {7,D}
4 Cd u0 {1,S} {6,D}
5 C u0 {2,D}
6 Cdd u0 {4,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)C=S(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 CS u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 C u0 {3,D}
7 S2d u0 {4,D}
8 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)C=S(Cds-Cdd-S2d)",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 CS u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 C u0 {3,D}
7 S2d u0 {4,D}
8 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)C=SCs",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)C=SCt",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 S2d u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)C=SCb",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 S2d u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-Cd)C=SC=S",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {6,D}
4 CS u0 {1,S} {7,D}
5 C u0 {2,D}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cd)C=S",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {7,D}
5 S2d u0 {2,D}
6 C u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cds)C=S",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {7,D}
5 S2d u0 {2,D}
6 Cd u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd)C=S",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {7,D}
5 S2d u0 {2,D}
6 Cdd u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-S2d)C=S",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 CS u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
8 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)(Cds-Cdd-Cd)C=S",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 Cdd u0 {1,D} {6,D}
4 CS u0 {1,S} {7,D}
5 Cdd u0 {2,D} {8,D}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
8 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsCbC=S",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cb u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsCtC=S",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Ct u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1204,
label = "Cds-CdsC=SCs",
group =
"""
1 * Cd u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 Cd u0 {1,D}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.44,4.73,4.94,5.14,5.48,5.75,6.24],'cal/(mol*K)'),
H298 = (10.34,'kcal/mol'),
S298 = (-11.67,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsC=S(Cds-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CS u0 {1,S} {6,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,D}
5 C u0 {3,D}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsC=S(Cds-Cds)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CS u0 {1,S} {6,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,D}
5 Cd u0 {3,D}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsC=S(Cds-Cdd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CS u0 {1,S} {6,D}
3 Cd u0 {1,S} {5,D}
4 Cd u0 {1,D}
5 Cdd u0 {3,D}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsC=S(Cds-Cdd-Cd)",
group =
"""
1 * Cd u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cd u0 {1,D}
6 S2d u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-CdsC=S(Cds-Cdd-S2d)",
group =
"""
1 * Cd u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 Cd u0 {1,D}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cds-(Cdd-S2d)C=SC=S",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cdd u0 {1,D} {5,D}
3 CS u0 {1,S} {6,D}
4 CS u0 {1,S} {7,D}
5 S2d u0 {2,D}
6 S2d u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1996,
label = "Cd-CdCs(CO)",
group =
"""
1 * Cd u0 {2,S} {3,S} {4,D}
2 CO u0 {1,S} {5,D}
3 Cs u0 {1,S}
4 Cd u0 {1,D}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([15.33,16.82,18.64,20.42,23.2,25,27.1],'J/(mol*K)','+|-',[5.66,5.66,5.66,5.66,5.66,5.66,5.66]),
H298 = (39,'kJ/mol','+|-',4.82),
S298 = (-51.26,'J/(mol*K)','+|-',6.6),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 2011,
label = "Cd-CdCs(CCO)",
group =
"""
1 * Cd u0 {2,S} {4,S} {5,D}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 Cs u0 {1,S}
5 Cd u0 {1,D}
6 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([22.68,24.05,24.63,25.07,25.64,25.84,25.7],'J/(mol*K)','+|-',[8,8,8,8,8,8,8]),
H298 = (41.6,'kJ/mol','+|-',6.82),
S298 = (-48.01,'J/(mol*K)','+|-',9.33),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1923,
label = "Cds-CNH",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 N u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1860,
label = "Cd-CdHN3s",
group =
"""
1 * Cd u0 {2,D} {5,S} {6,S}
2 Cd u0 {1,D} {3,S} {4,S}
3 R u0 {2,S}
4 R u0 {2,S}
5 H u0 {1,S}
6 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.7,6,7,7.7,8.8,9.5,10.6],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (2.2,'kcal/mol','+|-',1.4),
S298 = (7.1,'cal/(mol*K)','+|-',1.3),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1838,
label = "Cd-CdH(N5dOdOs)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D} {5,S} {6,S}
3 N5dc u0 {1,S} {7,D} {8,S}
4 H u0 {1,S}
5 R u0 {2,S}
6 R u0 {2,S}
7 O2d u0 {3,D}
8 O2s u0 {3,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.7,15.4,17.6,19.3,21.7,23.1,25],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (2,'kcal/mol','+|-',1.3),
S298 = (44.3,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1924,
label = "Cds-CCN",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 C u0 {1,D}
3 C u0 {1,S}
4 N u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1859,
label = "Cd-CdCsN3s",
group =
"""
1 * Cd u0 {2,D} {5,S} {6,S}
2 Cd u0 {1,D} {3,S} {4,S}
3 R u0 {2,S}
4 R u0 {2,S}
5 Cs u0 {1,S}
6 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.8,5,5.9,6.4,6.9,7.1,7.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (3.5,'kcal/mol','+|-',1.4),
S298 = (-14.1,'cal/(mol*K)','+|-',1.3),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1861,
label = "Cd-CdCs(N5dOdOs)",
group =
"""
1 * Cd u0 {2,D} {3,S} {4,S}
2 Cd u0 {1,D} {5,S} {6,S}
3 N5dc u0 {1,S} {7,D} {8,S}
4 Cs u0 {1,S}
5 R u0 {2,S}
6 R u0 {2,S}
7 O2d u0 {3,D}
8 O2s u0 {3,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.1,14.3,16.1,17.5,19.3,20.3,21.4],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (2.3,'kcal/mol','+|-',1.3),
S298 = (24,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-SsSs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CH",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 C u0 {1,S}
4 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1184,
label = "C=S-CsH",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.11,9.03,9.88,10.61,11.74,12.55,13.82],'cal/(mol*K)'),
H298 = (27.32,'kcal/mol'),
S298 = (37.56,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1185,
label = "C=S-CdsH",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cd u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.59,9.38,10.81,11.85,13.18,13.95,14.81],'cal/(mol*K)'),
H298 = (24.05,'kcal/mol'),
S298 = (34.35,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cd)H",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 H u0 {1,S}
5 [Cd,Cdd,CO] u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd)H",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 H u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-Cd)H",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 H u0 {1,S}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-S2d)H",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 H u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cds)H",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 H u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1186,
label = "C=S-CtH",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.46,8.91,10.01,10.83,11.98,12.74,13.87],'cal/(mol*K)'),
H298 = (30.83,'kcal/mol'),
S298 = (37.16,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1187,
label = "C=S-CbH",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.45,9.84,10.94,11.78,12.97,13.76,14.77],'cal/(mol*K)'),
H298 = (24.71,'kcal/mol'),
S298 = (34.15,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1188,
label = "C=S-C=SH",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 H u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.79,9.18,10.41,11.42,12.82,13.63,14.54],'cal/(mol*K)'),
H298 = (26.96,'kcal/mol'),
S298 = (35.65,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CC",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 C u0 {1,S}
4 C u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CbCds",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cb u0 {1,S}
4 Cd u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Cb(Cds-Cd)",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cb u0 {1,S}
5 [Cd,Cdd,CO] u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Cb(Cds-Cds)",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cb u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Cb(Cds-Cdd)",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cb u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Cb(Cds-Cdd-S2d)",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 Cb u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Cb(Cds-Cdd-Cd)",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 Cb u0 {1,S}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CtCt",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CbCb",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CdsCds",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cd u0 {1,S}
4 Cd u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cd)(Cds-Cd)",
group =
"""
1 * CS u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2d u0 {1,D}
5 [Cd,Cdd,CO] u0 {2,D}
6 [Cd,Cdd,CO] u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd)(Cds-Cds)",
group =
"""
1 * CS u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2d u0 {1,D}
5 Cdd u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-Cd)(Cds-Cds)",
group =
"""
1 * CS u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 S2d u0 {1,D}
6 Cd u0 {3,D}
7 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-S2d)(Cds-Cds)",
group =
"""
1 * CS u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {6,D}
4 Cdd u0 {2,D} {7,D}
5 S2d u0 {1,D}
6 Cd u0 {3,D}
7 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cds)(Cds-Cds)",
group =
"""
1 * CS u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2d u0 {1,D}
5 Cd u0 {2,D}
6 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * CS u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 S2d u0 {1,D}
5 Cdd u0 {2,D}
6 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * CS u0 {2,S} {3,S} {6,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {7,D}
5 Cdd u0 {3,D} {8,D}
6 S2d u0 {1,D}
7 C u0 {4,D}
8 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * CS u0 {2,S} {3,S} {6,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {7,D}
5 Cdd u0 {3,D} {8,D}
6 S2d u0 {1,D}
7 S2d u0 {4,D}
8 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-Cd)(Cds-Cdd-S2d)",
group =
"""
1 * CS u0 {2,S} {3,S} {6,D}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {7,D}
5 Cdd u0 {3,D} {8,D}
6 S2d u0 {1,D}
7 C u0 {4,D}
8 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CtCds",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Ct u0 {1,S}
4 Cd u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Ct(Cds-Cd)",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Ct u0 {1,S}
5 [Cd,Cdd,CO] u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Ct(Cds-Cds)",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Ct u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Ct(Cds-Cdd)",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Ct u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Ct(Cds-Cdd-Cd)",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 Ct u0 {1,S}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-Ct(Cds-Cdd-S2d)",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 Ct u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CbCt",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1190,
label = "C=S-CsCs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.7,7.44,8.14,8.72,9.52,9.98,10.51],'cal/(mol*K)'),
H298 = (27.2,'kcal/mol'),
S298 = (18,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1191,
label = "C=S-CdsCs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cd u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.79,9.21,10.13,10.71,11.25,11.42,11.35],'cal/(mol*K)'),
H298 = (26.19,'kcal/mol'),
S298 = (13.44,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cd)Cs",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cs u0 {1,S}
5 [Cd,Cdd,CO] u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cds)Cs",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cs u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd)Cs",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cs u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-S2d)Cs",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 Cs u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-Cd)Cs",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 Cs u0 {1,S}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1192,
label = "C=S-CtCs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.87,7.88,8.6,9.13,9.8,10.17,10.59],'cal/(mol*K)'),
H298 = (30.12,'kcal/mol'),
S298 = (17.46,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1193,
label = "C=S-CbCs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.02,9.02,9.75,10.23,10.75,10.96,11.04],'cal/(mol*K)'),
H298 = (26.6,'kcal/mol'),
S298 = (14.55,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1194,
label = "C=S-C=SCs",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cs u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.93,7.93,8.76,9.37,10.11,10.45,10.71],'cal/(mol*K)'),
H298 = (27.48,'kcal/mol'),
S298 = (16.58,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CtC=S",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Ct u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cd)C=S",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cd u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd)C=S",
group =
"""
1 * CS u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {6,D}
4 S2d u0 {1,D}
5 Cdd u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-Cd)C=S",
group =
"""
1 * CS u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {6,D}
5 S2d u0 {1,D}
6 C u0 {4,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-S2d)C=S",
group =
"""
1 * CS u0 {2,S} {3,S} {5,D}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {6,D}
5 S2d u0 {1,D}
6 S2d u0 {4,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cds)C=S",
group =
"""
1 * CS u0 {2,S} {3,S} {4,D}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {6,D}
4 S2d u0 {1,D}
5 Cd u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-C=SC=S",
group =
"""
1 * CS u0 {2,S} {3,S} {4,D}
2 CS u0 {1,S} {5,D}
3 CS u0 {1,S} {6,D}
4 S2d u0 {1,D}
5 S2d u0 {2,D}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CbC=S",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 Cb u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1183,
label = "C=S-HH",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([9.08,10.34,11.51,12.5,14.07,15.25,17.14],'cal/(mol*K)'),
H298 = (27.71,'kcal/mol'),
S298 = (56.51,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1189,
label = "C=S-SsH",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.38,9.78,10.83,11.66,12.86,13.71,14.87],'cal/(mol*K)'),
H298 = (21.55,'kcal/mol'),
S298 = (34.41,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1205,
label = "C=S-CSs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 C u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = u'C=S-CsSs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CbSs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CdsSs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cd)S2s",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 S2s u0 {1,S}
5 [Cd,Cdd,CO] u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cds)S2s",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 S2s u0 {1,S}
5 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd)S2s",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 Cd u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 S2s u0 {1,S}
5 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-Cd)S2s",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 S2s u0 {1,S}
6 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-(Cds-Cdd-S2d)S2s",
group =
"""
1 * CS u0 {2,S} {4,D} {5,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {6,D}
4 S2d u0 {1,D}
5 S2s u0 {1,S}
6 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-CtSs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1206,
label = "C=S-CsSs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.4,8.38,9.16,9.8,10.72,11.25,11.66],'cal/(mol*K)'),
H298 = (21.35,'kcal/mol'),
S298 = (14.52,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "C=S-C=SSs",
group =
"""
1 * CS u0 {2,S} {3,D} {4,S}
2 CS u0 {1,S} {5,D}
3 S2d u0 {1,D}
4 S2s u0 {1,S}
5 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1457,
label = "CS-OsH",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.48,5.3,6.09,6.82,8.05,8.99,10.37],'cal/(mol*K)'),
H298 = (2.85,'kcal/mol'),
S298 = (30.14,'cal/(mol*K)'),
),
shortDesc = u"""CAC 1d-HR calc""",
longDesc =
u"""
""",
)
entry(
index = 1458,
label = "CS-CsOs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 O2s u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.9,4.17,4.6,5.1,6.08,6.76,7.44],'cal/(mol*K)'),
H298 = (-1.32,'kcal/mol'),
S298 = (8.62,'cal/(mol*K)'),
),
shortDesc = u"""CAC 1d-HR calc""",
longDesc =
u"""
""",
)
entry(
index = 1459,
label = "CS-OsOs",
group =
"""
1 * CS u0 {2,D} {3,S} {4,S}
2 S2d u0 {1,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.08,3.59,3.9,4.03,3.99,3.75,3.23],'cal/(mol*K)'),
H298 = (-22.72,'kcal/mol'),
S298 = (2.67,'cal/(mol*K)'),
),
shortDesc = u"""CAC CBS-QB3 1Dhr calc""",
longDesc =
u"""
""",
)
entry(
index = 329,
label = "Cs",
group =
"""
1 * Cs u0
""",
thermo = u'Cs-CsCsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1919,
label = "Cs-NHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1800,
label = "Cs-N3sHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3s u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)'),
H298 = (-10.08,'kcal/mol'),
S298 = (30.41,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1920,
label = "Cs-N3dHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1870,
label = "Cs-(N3dCd)HHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S} {6,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 [Cd,Cdd] u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6,7.7,9.3,10.7,13.1,14.8,17.7],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (-5.7,'kcal/mol','+|-',1.3),
S298 = (30.4,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1801,
label = "Cs-(N3dN3d)HHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S} {6,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 N3d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6,7.8,9.4,10.8,13.1,14.8,17.6],'cal/(mol*K)','+|-',[0.6,0.6,0.6,0.6,0.6,0.6,0.6]),
H298 = (-9,'kcal/mol','+|-',0.8),
S298 = (30.2,'cal/(mol*K)','+|-',0.8),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1921,
label = "Cs-NCsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1802,
label = "Cs-N3sCsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3s u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.25,6.9,8.28,9.39,11.09,12.34,14.8],'cal/(mol*K)'),
H298 = (-6.6,'kcal/mol'),
S298 = (9.8,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1925,
label = "Cs-N3dCHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1805,
label = "Cs-(N3dN3d)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 N3d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.3,6.9,8.3,9.4,11.1,12.3,14.2],'cal/(mol*K)','+|-',[0.6,0.6,0.6,0.6,0.6,0.6,0.6]),
H298 = (-5.5,'kcal/mol','+|-',0.8),
S298 = (9.4,'cal/(mol*K)','+|-',0.8),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1869,
label = "Cs-(N3dOd)CHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.8,13.6,15.2,16.7,18.9,20.5,23],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (21.4,'kcal/mol','+|-',1.3),
S298 = (44.3,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1871,
label = "Cs-(N3dCd)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.3,7.2,8.7,9.8,11.6,12.8,14.7],'cal/(mol*K)','+|-',[1.2,1.2,1.2,1.2,1.2,1.2,1.2]),
H298 = (-2.9,'kcal/mol','+|-',1.7),
S298 = (8.6,'cal/(mol*K)','+|-',1.6),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1926,
label = "Cs-N5dCsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N5dc u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1841,
label = "Cs-(N5dOdOs)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N5dc u0 {1,S} {6,D} {7,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2s u0 {2,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.9,15.8,18.3,20.3,23.3,25.4,28.3],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (-14.8,'kcal/mol','+|-',1.3),
S298 = (48.9,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1927,
label = "Cs-NCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1803,
label = "Cs-N3sCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3s u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.67,6.32,7.64,8.39,9.56,10.23,11.905],'cal/(mol*K)'),
H298 = (-5.2,'kcal/mol'),
S298 = (-11.7,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1928,
label = "Cs-N3dCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1868,
label = "Cs-(N3dOd)CsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.2,12.7,14,15.1,16.8,17.9,19.5],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (23.4,'kcal/mol','+|-',1.3),
S298 = (23.1,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1929,
label = "Cs-N5dCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N5dc u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1842,
label = "Cs-(N5dOdOs)CsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N5dc u0 {1,S} {6,D} {7,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2s u0 {2,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([13.6,16.1,18.1,19.6,21.8,23.2,25.1],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (-13.9,'kcal/mol','+|-',1.3),
S298 = (27.5,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1930,
label = "Cs-NCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1804,
label = "Cs-N3sCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3s u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.35,6.16,7.31,7.91,8.49,8.5,8.525],'cal/(mol*K)'),
H298 = (-3.2,'kcal/mol'),
S298 = (-34.1,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1806,
label = "Cs-(N3dN3d)CsCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 N3d u0 {1,S} {3,D}
3 N3d u0 {2,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (-3.3,'kcal/mol'),
S298 = (-11.7,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1807,
label = "Cs-(N3dN3d)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 N3d u0 {1,S} {3,D}
3 N3d u0 {2,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (-1.9,'kcal/mol'),
S298 = (-34.7,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1931,
label = "Cs-N3dCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1867,
label = "Cs-(N3dOd)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N3d u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.2,13.3,14,14.5,15.3,15.7,16.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (24.1,'kcal/mol','+|-',1.3),
S298 = (1.2,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1932,
label = "Cs-N5dCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N5dc u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1843,
label = "Cs-(N5dOdOs)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N5dc u0 {1,S} {6,D} {7,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
7 O2s u0 {2,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([13.5,16,17.8,18.9,20.3,21.1,21.9],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (-12.7,'kcal/mol','+|-',1.3),
S298 = (5.2,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1933,
label = "Cs-NNCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N u0 {1,S}
3 N u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1934,
label = "Cs-N5dN5dCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N5dc u0 {1,S}
3 N5dc u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1933,
label = "Cs-NNCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N u0 {1,S}
3 N u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1846,
label = "Cs-(N5dOdOs)(N5dOdOs)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 N5dc u0 {1,S} {6,D} {7,S}
3 N5dc u0 {1,S} {8,D} {9,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2s u0 {2,S}
8 O2d u0 {3,D}
9 O2s u0 {3,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (-14.9,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 330,
label = "Cs-HHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 H u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.43,9.84,11.14,12.41,15,17.25,20.63],'cal/(mol*K)','+|-',[0.06,0.06,0.06,0.06,0.06,0.06,0.06]),
H298 = (-17.9,'kcal/mol','+|-',0.1),
S298 = (49.41,'cal/(mol*K)','+|-',0.05),
),
shortDesc = u"""CHEMKIN DATABASE S(group) = S(CH4) + Rln(12)""",
longDesc =
u"""
""",
)
entry(
index = 331,
label = "Cs-CHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CsHHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 332,
label = "Cs-CsHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]),
H298 = (-10.2,'kcal/mol','+|-',0.12),
S298 = (30.41,'cal/(mol*K)','+|-',0.08),
),
shortDesc = u"""Cs-CsHHH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 333,
label = "Cs-CdsHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)HHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 334,
label = "Cs-(Cds-O2d)HHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([25.31,32.07,38.44,44.06,53.36,60.63,72.47],'J/(mol*K)'),
H298 = (-42.9,'kJ/mol'),
S298 = (127.12,'J/(mol*K)'),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 335,
label = "Cs-(Cds-Cd)HHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)HHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 336,
label = "Cs-(Cds-Cds)HHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)','+|-',[0.04,0.04,0.04,0.04,0.04,0.04,0.04]),
H298 = (-10.2,'kcal/mol','+|-',0.08),
S298 = (30.41,'cal/(mol*K)','+|-',0.04),
),
shortDesc = u"""Cs-CdHHH BENSON (Assigned Cs-CsHHH)""",
longDesc =
u"""
""",
)
entry(
index = 337,
label = "Cs-(Cds-Cdd)HHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)HHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 338,
label = "Cs-(Cds-Cdd-O2d)HHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([25.31,32.07,38.44,44.06,53.36,60.63,72.47],'J/(mol*K)'),
H298 = (-42.9,'kJ/mol'),
S298 = (127.12,'J/(mol*K)'),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)HHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 339,
label = "Cs-(Cds-Cdd-Cd)HHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)HHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1862,
label = "Cs-(CdN3d)HHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 N3d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.2,7.8,9.4,10.8,13,14.8,17.6],'cal/(mol*K)'),
H298 = (-10.2,'kcal/mol'),
S298 = (30.4,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 340,
label = "Cs-CtHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]),
H298 = (-10.2,'kcal/mol','+|-',0.15),
S298 = (30.41,'cal/(mol*K)','+|-',0.08),
),
shortDesc = u"""Cs-CtHHH BENSON (Assigned Cs-CsHHH)""",
longDesc =
u"""
""",
)
entry(
index = 1863,
label = "Cs-(CtN3t)HHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S} {6,T}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 N3t u0 {2,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.5,14.6,16.6,18.3,21.1,23.4,26.9],'cal/(mol*K)','+|-',[1.3,1.3,1.3,1.3,1.3,1.3,1.3]),
H298 = (17.7,'kcal/mol','+|-',1.9),
S298 = (60.2,'cal/(mol*K)','+|-',1.7),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 341,
label = "Cs-CbHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.19,7.84,9.4,10.79,13.02,14.77,17.58],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-10.2,'kcal/mol','+|-',0.18),
S298 = (30.41,'cal/(mol*K)','+|-',0.14),
),
shortDesc = u"""Cs-CbHHH BENSON (Assigned Cs-CsHHH)""",
longDesc =
u"""
""",
)
entry(
index = 1176,
label = "Cs-C=SHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.96,7.6,9.13,10.49,12.72,14.46,17.28],'cal/(mol*K)'),
H298 = (-10.25,'kcal/mol'),
S298 = (30.4,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 342,
label = "Cs-OsHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 O2s u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([25.31,32.07,38.44,44.06,53.36,60.63,72.47],'J/(mol*K)'),
H298 = (-42.9,'kJ/mol'),
S298 = (127.12,'J/(mol*K)'),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 343,
label = "Cs-OsOsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 O2s u0 {1,S}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([27.35,43.68,53.55,58.15,60.86,61.66,63.53],'J/(mol*K)','+|-',[5.77,5.77,5.77,5.77,5.77,5.77,5.77]),
H298 = (-67.5,'kJ/mol','+|-',4.92),
S298 = (17.89,'J/(mol*K)','+|-',6.74),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 344,
label = "Cs-OsOsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 O2s u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.54,6,7.17,8.05,9.31,10.05,10.05],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-21.23,'kcal/mol','+|-',0.2),
S298 = (-12.07,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cs-OOOH BOZZELLI del C/C2/O - C/C3/O, series !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1451,
label = "Cs-OsSsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 O2s u0 {1,S}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.48,9.54,11,11.91,12.85,13.54,14.93],'cal/(mol*K)'),
H298 = (-11.58,'kcal/mol'),
S298 = (4.58,'cal/(mol*K)'),
),
shortDesc = u"""CAC CBS-QB3 1DHR CAC""",
longDesc =
u"""
""",
)
entry(
index = 1464,
label = "Cs-OsOsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 O2s u0 {1,S}
3 O2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.36,8.72,10.13,10.88,11.56,11.91,12.53],'cal/(mol*K)'),
H298 = (-19.72,'kcal/mol'),
S298 = (-13.26,'cal/(mol*K)'),
),
shortDesc = u"""CAC calc 1D-HR""",
longDesc =
u"""
""",
)
entry(
index = 1162,
label = "Cs-SsHHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 S2s u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.96,7.6,9.13,10.49,12.72,14.46,17.28],'cal/(mol*K)'),
H298 = (-10.25,'kcal/mol'),
S298 = (30.4,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1167,
label = "Cs-SsSsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 S2s u0 {1,S}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.37,9.7,10.52,11.13,12.16,13.01,14.43],'cal/(mol*K)'),
H298 = (-6.21,'kcal/mol'),
S298 = (6.14,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1201,
label = "Cs-SsSsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 S2s u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.84,9.14,10.24,10.73,11.12,11.33,11.57],'cal/(mol*K)'),
H298 = (-2.78,'kcal/mol'),
S298 = (-15.38,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 345,
label = "Cs-CCHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CsCsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 346,
label = "Cs-CsCsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.5,6.95,8.25,9.35,11.07,12.34,14.25],'cal/(mol*K)','+|-',[0.04,0.04,0.04,0.04,0.04,0.04,0.04]),
H298 = (-4.93,'kcal/mol','+|-',0.05),
S298 = (9.42,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CsCsHH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 347,
label = "Cs-CdsCsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 348,
label = "Cs-(Cds-O2d)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([26.91,30.8,34.98,38.91,45.56,50.73,58.93],'J/(mol*K)','+|-',[1.53,1.53,1.53,1.53,1.53,1.53,1.53]),
H298 = (-21.5,'kJ/mol','+|-',1.3),
S298 = (40.32,'J/(mol*K)','+|-',1.78),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 349,
label = "Cs-(Cds-Cd)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 350,
label = "Cs-(Cds-Cds)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.12,6.86,8.32,9.49,11.22,12.48,14.36],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-4.76,'kcal/mol','+|-',0.16),
S298 = (9.8,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cs-CdCsHH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 351,
label = "Cs-(Cds-Cdd)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 352,
label = "Cs-(Cds-Cdd-O2d)CsHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.35,6.83,8.25,9.45,11.19,12.46,14.34],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-5.723,'kcal/mol','+|-',0.16),
S298 = (9.37,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{C/C/H2/CCO} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CsHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 353,
label = "Cs-(Cds-Cdd-Cd)CsHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1864,
label = "Cs-(CdN3d)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D} {7,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 N3d u0 {2,D}
7 R u0 {2,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.5,6.9,8.1,9.2,10.9,12.2,14.1],'cal/(mol*K)','+|-',[1.2,1.2,1.2,1.2,1.2,1.2,1.2]),
H298 = (-5.1,'kcal/mol','+|-',1.7),
S298 = (10.1,'cal/(mol*K)','+|-',1.6),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 354,
label = "Cs-CdsCdsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 355,
label = "Cs-(Cds-O2d)(Cds-O2d)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([27.77,32.81,37.1,40.67,46.39,50.85,58.25],'J/(mol*K)','+|-',[4.19,4.19,4.19,4.19,4.19,4.19,4.19]),
H298 = (-10,'kJ/mol','+|-',3.57),
S298 = (40.1,'J/(mol*K)','+|-',4.88),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 356,
label = "Cs-(Cds-O2d)(Cds-Cd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([24.94,31.41,36.47,40.49,46.72,51.49,59.29],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]),
H298 = (-16.9,'kJ/mol','+|-',2.85),
S298 = (40.18,'J/(mol*K)','+|-',3.9),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 357,
label = "Cs-(Cds-O2d)(Cds-Cds)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.75,7.11,8.92,10.32,12.16,13.61,13.61],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-3.8,'kcal/mol','+|-',0.16),
S298 = (6.31,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cs-COCdHH BENSON Hf, Mopac =3D S,Cp nov99 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 358,
label = "Cs-(Cds-O2d)(Cds-Cdd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 359,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 H u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 360,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 H u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 361,
label = "Cs-(Cds-Cd)(Cds-Cd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 362,
label = "Cs-(Cds-Cds)(Cds-Cds)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.7,6.8,8.4,9.6,11.3,12.6,14.4],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-4.29,'kcal/mol','+|-',0.16),
S298 = (10.2,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cs-CdCdHH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 363,
label = "Cs-(Cds-Cdd)(Cds-Cds)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 H u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 365,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 H u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 366,
label = "Cs-(Cds-Cdd)(Cds-Cdd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 367,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 H u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.68,8.28,9.58,10.61,12.04,13.13,14.87],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-5.301,'kcal/mol','+|-',0.16),
S298 = (7.18,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{C/H2/CCO2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 368,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 H u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-Cd(CCO)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 H u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 H u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 369,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 H u0 {1,S}
7 H u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 2010,
label = "Cs-Cd(CCO)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 H u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([25.85,31.99,37.06,41.14,47.42,52.15,59.73],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]),
H298 = (-22.2,'kJ/mol','+|-',5.9),
S298 = (37.92,'J/(mol*K)','+|-',8.08),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 370,
label = "Cs-CtCsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.95,6.56,7.93,9.08,10.86,12.19,14.2],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]),
H298 = (-4.73,'kcal/mol','+|-',0.28),
S298 = (10.3,'cal/(mol*K)','+|-',0.07),
),
shortDesc = u"""Cs-CtCsHH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 1832,
label = "Cs-(CtN3t)CsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S} {6,T}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 N3t u0 {2,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.3,13.5,15.3,16.8,19.2,20.9,23.5],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (22.9,'kcal/mol','+|-',1.3),
S298 = (39.8,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 371,
label = "Cs-CtCdsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 372,
label = "Cs-(Cds-O2d)CtHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.85,6.22,8.01,9.43,11.29,12.76,12.76],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]),
H298 = (-5.4,'kcal/mol','+|-',0.28),
S298 = (7.68,'cal/(mol*K)','+|-',0.07),
),
shortDesc = u"""Cs-COCtHH BENSON Hf, Mopac =3D S,Cp nov99 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 373,
label = "Cs-(Cds-Cd)CtHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CtHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 374,
label = "Cs-(Cds-Cds)CtHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.4,6.33,7.9,9.16,10.93,12.29,13.43],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]),
H298 = (-3.49,'kcal/mol','+|-',0.28),
S298 = (9.31,'cal/(mol*K)','+|-',0.07),
),
shortDesc = u"""Cs-CtCdHH RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 375,
label = "Cs-(Cds-Cdd)CtHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 376,
label = "Cs-(Cds-Cdd-O2d)CtHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-Cd(CCO)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 377,
label = "Cs-(Cds-Cdd-Cd)CtHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 378,
label = "Cs-CtCtHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4,6.07,7.71,9.03,10.88,12.3,12.48],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]),
H298 = (-0.82,'kcal/mol','+|-',0.28),
S298 = (10.04,'cal/(mol*K)','+|-',0.07),
),
shortDesc = u"""Cs-CtCtHH RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 379,
label = "Cs-CbCsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.84,7.61,8.98,10.01,11.49,12.54,13.76],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-4.86,'kcal/mol','+|-',0.2),
S298 = (9.34,'cal/(mol*K)','+|-',0.19),
),
shortDesc = u"""Cs-CbCsHH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 380,
label = "Cs-CbCdsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 381,
label = "Cs-(Cds-O2d)CbHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.38,7.59,9.25,10.51,12.19,13.52,13.52],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-5.4,'kcal/mol','+|-',0.2),
S298 = (5.89,'cal/(mol*K)','+|-',0.19),
),
shortDesc = u"""Cs-COCbHH BENSON Hf, Mopac =3D S,Cp nov99 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 382,
label = "Cs-(Cds-Cd)CbHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 383,
label = "Cs-(Cds-Cds)CbHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.51,6.76,8.61,10.01,11.97,13.4,15.47],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]),
H298 = (-4.29,'kcal/mol','+|-',0.2),
S298 = (2,'cal/(mol*K)','+|-',0.19),
),
shortDesc = u"""Cs-CbCdHH Hf=Stein S,Cp=3D mopac nov99""",
longDesc =
u"""
""",
)
entry(
index = 384,
label = "Cs-(Cds-Cdd)CbHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 385,
label = "Cs-(Cds-Cdd-O2d)CbHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-Cd(CCO)HH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 386,
label = "Cs-(Cds-Cdd-Cd)CbHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 387,
label = "Cs-CbCtHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.28,6.43,8.16,9.5,11.36,12.74,13.7],'cal/(mol*K)','+|-',[0.08,0.08,0.08,0.08,0.08,0.08,0.08]),
H298 = (-4.29,'kcal/mol','+|-',0.28),
S298 = (9.84,'cal/(mol*K)','+|-',0.07),
),
shortDesc = u"""Cs-CbCtHH Hf=Stein S,Cp=3D mopac nov99""",
longDesc =
u"""
""",
)
entry(
index = 388,
label = "Cs-CbCbHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.67,7.7,9.31,10.52,12.21,13.47,15.11],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]),
H298 = (-4.29,'kcal/mol','+|-',0.2),
S298 = (8.07,'cal/(mol*K)','+|-',0.19),
),
shortDesc = u"""Cs-CbCbHH Hf=3Dbsn/Cs/Cd2/H2 S,Cp=3D mopac nov99""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1177,
label = "Cs-C=SCsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.23,6.82,8.16,9.27,10.96,12.2,14.13],'cal/(mol*K)'),
H298 = (-4.89,'kcal/mol'),
S298 = (9.83,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 H u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 H u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)HH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 H u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 389,
label = "Cs-CCCH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 C u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CsCsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 390,
label = "Cs-CsCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.54,6,7.17,8.05,9.31,10.05,11.17],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (-1.9,'kcal/mol','+|-',0.15),
S298 = (-12.07,'cal/(mol*K)','+|-',0.07),
),
shortDesc = u"""Cs-CsCsCsH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 391,
label = "Cs-CdsCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 392,
label = "Cs-(Cds-O2d)CsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([23.68,27.86,31.26,34,38.07,41,45.46],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]),
H298 = (-5.4,'kJ/mol','+|-',2.85),
S298 = (-47.41,'J/(mol*K)','+|-',3.9),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 393,
label = "Cs-(Cds-Cd)CsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 394,
label = "Cs-(Cds-Cds)CsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.16,5.91,7.34,8.19,9.46,10.19,11.28],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-1.48,'kcal/mol','+|-',0.27),
S298 = (-11.69,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CdCsCsH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 395,
label = "Cs-(Cds-Cdd)CsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 396,
label = "Cs-(Cds-Cdd-O2d)CsCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([21.23,27.55,32.36,35.85,40.37,43.16,46.94],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]),
H298 = (-11.1,'kJ/mol','+|-',5.9),
S298 = (-47.59,'J/(mol*K)','+|-',8.08),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CsCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 397,
label = "Cs-(Cds-Cdd-Cd)CsCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.16,5.91,7.34,8.19,9.46,10.19,11.28],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-1.48,'kcal/mol','+|-',0.27),
S298 = (-11.69,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CdCsCsH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 1865,
label = "Cs-(CdN3d)CsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D} {7,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 N3d u0 {2,D}
7 R u0 {2,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5,6.5,7.5,8.2,9.3,9.9,10.9],'cal/(mol*K)','+|-',[1.2,1.2,1.2,1.2,1.2,1.2,1.2]),
H298 = (-1.6,'kcal/mol','+|-',1.7),
S298 = (-11.2,'cal/(mol*K)','+|-',1.6),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 398,
label = "Cs-CtCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,5.61,6.85,7.78,9.1,9.9,11.12],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-1.72,'kcal/mol','+|-',0.27),
S298 = (-11.19,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CtCsCsH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 1833,
label = "Cs-(CtN3t)CsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S} {6,T}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 N3t u0 {2,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11,12.7,14.1,15.4,17.3,18.6,21.85],'cal/(mol*K)'),
H298 = (25.8,'kcal/mol'),
S298 = (19.8,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 399,
label = "Cs-CbCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.88,6.66,7.9,8.75,9.73,10.25,10.68],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-0.98,'kcal/mol','+|-',0.27),
S298 = (-12.15,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CbCsCsH BENSON""",
longDesc =
u"""
""",
)
entry(
index = 400,
label = "Cs-CdsCdsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 401,
label = "Cs-(Cds-O2d)(Cds-O2d)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CsCsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 402,
label = "Cs-(Cds-O2d)(Cds-Cd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([29.32,32.99,35.49,37.28,39.75,41.6,44.96],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]),
H298 = (-2.2,'kJ/mol','+|-',2.85),
S298 = (-50.47,'J/(mol*K)','+|-',3.9),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 403,
label = "Cs-(Cds-O2d)(Cds-Cds)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 404,
label = "Cs-(Cds-O2d)(Cds-Cdd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 405,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 406,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 407,
label = "Cs-(Cds-Cd)(Cds-Cd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 408,
label = "Cs-(Cds-Cds)(Cds-Cds)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.28,6.54,7.67,8.48,9.45,10.18,11.24],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-1.1,'kcal/mol','+|-',0.27),
S298 = (-13.03,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CdCdCsH RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 409,
label = "Cs-(Cds-Cdd)(Cds-Cds)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 411,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 412,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 413,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([21.19,28,33.91,38.75,46.07,51.36,59.45],'J/(mol*K)','+|-',[3.46,3.46,3.46,3.46,3.46,3.46,3.46]),
H298 = (-21.1,'kJ/mol','+|-',2.95),
S298 = (40.95,'J/(mol*K)','+|-',4.04),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 414,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-CsCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 415,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 H u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 2008,
label = "Cs-CsCd(CCO)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([24.45,31.59,36.01,38.8,42.13,44.21,47.25],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]),
H298 = (-10.4,'kJ/mol','+|-',5.9),
S298 = (-54.03,'J/(mol*K)','+|-',8.08),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 416,
label = "Cs-CtCdsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 417,
label = "Cs-(Cds-O2d)CtCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 418,
label = "Cs-(Cds-Cd)CtCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CtCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 419,
label = "Cs-(Cds-Cds)CtCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.55,7.21,8.39,9.17,10,10.61,10.51],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-6.9,'kcal/mol','+|-',0.27),
S298 = (-13.48,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CtCdCsH RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 420,
label = "Cs-(Cds-Cdd)CtCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 421,
label = "Cs-(Cds-Cdd-O2d)CtCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CsCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 422,
label = "Cs-(Cds-Cdd-Cd)CtCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 423,
label = "Cs-CbCdsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 424,
label = "Cs-(Cds-O2d)CbCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 425,
label = "Cs-(Cds-Cd)CbCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 426,
label = "Cs-(Cds-Cds)CbCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.5,6.57,8.07,8.89,9.88,10.39,10.79],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-1.56,'kcal/mol','+|-',0.27),
S298 = (-11.77,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CbCdCsH BOZZELLI =3D Cs/Cs2/Cd/H + (Cs/Cs2/Cb/H - Cs/Cs3/H)""",
longDesc =
u"""
""",
)
entry(
index = 427,
label = "Cs-(Cds-Cdd)CbCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 428,
label = "Cs-(Cds-Cdd-O2d)CbCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CsCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 429,
label = "Cs-(Cds-Cdd-Cd)CbCsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 430,
label = "Cs-CtCtCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.27,5.32,6.9,8.03,9.33,10.21,9.38],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (1.72,'kcal/mol','+|-',0.27),
S298 = (-11.61,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CtCtCsH RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 431,
label = "Cs-CbCtCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.33,6.27,7.58,8.48,9.52,10.1,10.63],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-1.55,'kcal/mol','+|-',0.27),
S298 = (-11.65,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CbCtCsH BOZZELLI =3D Cs/Cs2/Cb/H + (Cs/Cs2/Ct/H - Cs/Cs3/H)""",
longDesc =
u"""
""",
)
entry(
index = 432,
label = "Cs-CbCbCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.22,7.32,8.63,8.45,10.15,10.45,10.89],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-1.06,'kcal/mol','+|-',0.27),
S298 = (-12.23,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CbCbCsCs BOZZELLI =3D Cs/Cs2/Cb/H + (Cs/Cs2/Cb/H - Cs/Cs3/H)""",
longDesc =
u"""
""",
)
entry(
index = 433,
label = "Cs-CdsCdsCdsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 434,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-CsCsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 435,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 H u0 {1,S}
6 C u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 436,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 H u0 {1,S}
6 Cd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 437,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 H u0 {1,S}
6 Cdd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 438,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 439,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 440,
label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 H u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([29.26,34.41,37.4,39.22,41.43,43.04,46.12],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]),
H298 = (2.9,'kJ/mol','+|-',2.85),
S298 = (-53.2,'J/(mol*K)','+|-',3.9),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 441,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsCsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 442,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 443,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 444,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 445,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 446,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 447,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 448,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 449,
label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 H u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 450,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.51,5.96,7.13,7.98,9.06,9.9,11.23],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (0.41,'kcal/mol','+|-',0.27),
S298 = (-11.82,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CdCdCdH RAMAN & GREEN JPC 2002""",
longDesc =
u"""
""",
)
entry(
index = 451,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 453,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 454,
label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 455,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 456,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-CdCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 457,
label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 Cd u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 458,
label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 459,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 H u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 460,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 H u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 461,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 H u0 {1,S}
9 O2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-CdCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 H u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 H u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 H u0 {1,S}
9 S2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 462,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 H u0 {1,S}
9 C u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 2009,
label = "Cs-CdCd(CCO)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([27.62,35.4,39.24,41.25,43.4,44.87,47.43],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]),
H298 = (-6.8,'kJ/mol','+|-',5.9),
S298 = (-55.37,'J/(mol*K)','+|-',8.08),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 463,
label = "Cs-CtCdsCdsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 464,
label = "Cs-(Cds-O2d)(Cds-O2d)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 465,
label = "Cs-(Cds-O2d)(Cds-Cd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 466,
label = "Cs-(Cds-O2d)(Cds-Cds)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 467,
label = "Cs-(Cds-O2d)(Cds-Cdd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 468,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 469,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 470,
label = "Cs-(Cds-Cd)(Cds-Cd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 471,
label = "Cs-(Cds-Cds)(Cds-Cds)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.68,7.85,8.62,9.16,9.81,10.42,10.49],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (1.88,'kcal/mol','+|-',0.27),
S298 = (-13.75,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CtCdCdH RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 472,
label = "Cs-(Cds-Cdd)(Cds-Cds)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 473,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-CdCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 474,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 475,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 476,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 477,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 478,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 H u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 479,
label = "Cs-CbCdsCdsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 480,
label = "Cs-(Cds-O2d)(Cds-O2d)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 481,
label = "Cs-(Cds-O2d)(Cds-Cd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 482,
label = "Cs-(Cds-O2d)(Cds-Cds)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 483,
label = "Cs-(Cds-O2d)(Cds-Cdd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 484,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 485,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 486,
label = "Cs-(Cds-Cd)(Cds-Cd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 487,
label = "Cs-(Cds-Cds)(Cds-Cds)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.12,6.51,8.24,9,10.03,10.53,10.89],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-1.39,'kcal/mol','+|-',0.27),
S298 = (-11.39,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CbCdCdH BOZZELLI =3D Cs/Cs/Cd2/H + (Cs/Cs2/Cb/H - Cs/Cs3/H)""",
longDesc =
u"""
""",
)
entry(
index = 488,
label = "Cs-(Cds-Cdd)(Cds-Cds)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 489,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-CdCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 490,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 491,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 492,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 493,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 494,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 H u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 495,
label = "Cs-CtCtCdsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CtCt(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 496,
label = "Cs-CtCt(Cds-O2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 497,
label = "Cs-CtCt(Cds-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-CtCt(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 498,
label = "Cs-CtCt(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.58,5.68,7.11,8.12,9.27,10.13,9.44],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (4.73,'kcal/mol','+|-',0.27),
S298 = (-11.46,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CtCtCdH RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 499,
label = "Cs-CtCt(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-CtCt(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 500,
label = "Cs-CtCt(Cds-Cdd-O2d)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CdCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCt(Cds-Cdd-S2d)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 501,
label = "Cs-CtCt(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-CtCt(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 502,
label = "Cs-CbCtCdsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CbCt(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 503,
label = "Cs-CbCt(Cds-O2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 504,
label = "Cs-CbCt(Cds-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-CbCt(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 505,
label = "Cs-CbCt(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 506,
label = "Cs-CbCt(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-CbCt(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 507,
label = "Cs-CbCt(Cds-Cdd-O2d)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCt(Cds-Cdd-S2d)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 508,
label = "Cs-CbCt(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-CbCt(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 509,
label = "Cs-CbCbCdsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CbCb(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 510,
label = "Cs-CbCb(Cds-O2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCb(Cds-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 511,
label = "Cs-CbCb(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 512,
label = "Cs-CbCb(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-CbCb(Cds-Cdd-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 513,
label = "Cs-CbCb(Cds-Cdd-O2d)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CdCd(CCO)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCb(Cds-Cdd-S2d)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 514,
label = "Cs-CbCb(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-CbCb(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 515,
label = "Cs-CtCtCtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.03,5.27,6.78,7.88,9.14,10.08,8.47],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (10.11,'kcal/mol','+|-',0.27),
S298 = (-10.46,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CtCtCtH RAMAN & GREEN JPCA 2002, 106, 11141-11149""",
longDesc =
u"""
""",
)
entry(
index = 516,
label = "Cs-CbCtCtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CtCt(Cds-Cds)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 517,
label = "Cs-CbCbCtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 518,
label = "Cs-CbCbCbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.56,7.98,9.36,10.15,10.57,10.65,9.7],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-0.34,'kcal/mol','+|-',0.27),
S298 = (-12.31,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CbCbCbH BOZZELLI =3D Cs/Cs/Cb2/H + (Cs/Cs2/Cb/H - Cs/Cs3/H)""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)(Cds-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 H u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCtH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CbH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCtC=SH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=SH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 H u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1178,
label = "Cs-C=SCsCsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.78,6.25,7.44,8.35,9.57,10.31,11.2],'cal/(mol*K)'),
H298 = (-0.78,'kcal/mol'),
S298 = (-11.46,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCtC=SH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCbC=SH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 H u0 {1,S}
6 C u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cds)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 H u0 {1,S}
6 Cd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 H u0 {1,S}
6 Cdd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-S2d)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-Cd)H",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 519,
label = "Cs-CCCC",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 C u0 {1,S}
5 C u0 {1,S}
""",
thermo = u'Cs-CsCsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 520,
label = "Cs-CsCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.37,6.13,7.36,8.12,8.77,8.76,8.12],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (0.5,'kcal/mol','+|-',0.27),
S298 = (-35.1,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CsCsCsCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 521,
label = "Cs-CdsCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 522,
label = "Cs-(Cds-O2d)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([22.68,27.48,30.12,31.51,32.36,32.39,32.42],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]),
H298 = (4.6,'kJ/mol','+|-',2.85),
S298 = (-140.94,'J/(mol*K)','+|-',3.9),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 523,
label = "Cs-(Cds-Cd)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 524,
label = "Cs-(Cds-Cds)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,6.04,7.43,8.26,8.92,8.96,8.23],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (1.68,'kcal/mol','+|-',0.27),
S298 = (-34.72,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CdCsCsCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 525,
label = "Cs-(Cds-Cdd)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 526,
label = "Cs-(Cds-Cdd-O2d)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([20.63,27.65,31.98,34.41,36.16,36.25,35.2],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]),
H298 = (-4.5,'kJ/mol','+|-',5.9),
S298 = (-144.08,'J/(mol*K)','+|-',8.08),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 527,
label = "Cs-(Cds-Cdd-Cd)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1866,
label = "Cs-(CdN3d)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D} {7,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 N3d u0 {2,D}
7 R u0 {2,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.3,6.6,7.3,7.5,7.8,7.8,7.7],'cal/(mol*K)','+|-',[1.2,1.2,1.2,1.2,1.2,1.2,1.2]),
H298 = (0.6,'kcal/mol','+|-',1.7),
S298 = (-33.5,'cal/(mol*K)','+|-',1.6),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 528,
label = "Cs-CtCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.37,6.79,8.09,8.78,9.19,8.96,7.63],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (2.81,'kcal/mol','+|-',0.27),
S298 = (-35.18,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""Cs-CtCsCsCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 1834,
label = "Cs-(CtN3t)CsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S} {6,T}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 N3t u0 {2,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.4,13.4,14.6,15.3,16.3,16.7,17.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (28.3,'kcal/mol','+|-',1.3),
S298 = (-3,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 529,
label = "Cs-CbCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.37,6.79,8.09,8.78,9.19,8.96,7.63],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (2.81,'kcal/mol','+|-',0.26),
S298 = (-35.18,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCsCsCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 530,
label = "Cs-CdsCdsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 531,
label = "Cs-(Cds-O2d)(Cds-O2d)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([33.76,33.42,32.6,31.91,31.01,30.55,30.35],'J/(mol*K)','+|-',[5.08,5.08,5.08,5.08,5.08,5.08,5.08]),
H298 = (14.9,'kJ/mol','+|-',4.33),
S298 = (-146.69,'J/(mol*K)','+|-',5.92),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 532,
label = "Cs-(Cds-O2d)(Cds-Cd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([26.01,30.13,32.44,33.51,33.75,33.26,32.55],'J/(mol*K)','+|-',[3.34,3.34,3.34,3.34,3.34,3.34,3.34]),
H298 = (9.8,'kJ/mol','+|-',2.85),
S298 = (-146.74,'J/(mol*K)','+|-',3.9),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 533,
label = "Cs-(Cds-O2d)(Cds-Cds)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 534,
label = "Cs-(Cds-O2d)(Cds-Cdd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 535,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 536,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 537,
label = "Cs-(Cds-Cd)(Cds-Cd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 538,
label = "Cs-(Cds-Cds)(Cds-Cds)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,6.04,7.43,8.26,8.92,8.96,8.23],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (1.68,'kcal/mol','+|-',0.26),
S298 = (-34.72,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CdCdCsCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 539,
label = "Cs-(Cds-Cdd)(Cds-Cds)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 541,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 542,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 543,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.73,8.1,9.02,9.53,9.66,9.52,8.93],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (-2.987,'kcal/mol','+|-',0.26),
S298 = (-36.46,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""{C/C2/CCO2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 544,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-CsCsCd(CCO)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 545,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 Cs u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 2007,
label = "Cs-CsCsCd(CCO)",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([25.48,31.89,35.19,36.68,37.19,36.66,34.96],'J/(mol*K)','+|-',[6.93,6.93,6.93,6.93,6.93,6.93,6.93]),
H298 = (2.9,'kJ/mol','+|-',5.9),
S298 = (-144.6,'J/(mol*K)','+|-',8.08),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 546,
label = "Cs-CtCdsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 547,
label = "Cs-(Cds-O2d)CtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 548,
label = "Cs-(Cds-Cd)CtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CtCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 549,
label = "Cs-(Cds-Cds)CtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,6.7,8.16,8.92,9.34,9.16,7.14],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (2.99,'kcal/mol','+|-',0.26),
S298 = (-34.8,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CtCdCsCs BOZZELLI =3D Cs/Cs3/Cd + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 550,
label = "Cs-(Cds-Cdd)CtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 551,
label = "Cs-(Cds-Cdd-O2d)CtCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CsCsCd(CCO)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 552,
label = "Cs-(Cds-Cdd-Cd)CtCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 553,
label = "Cs-CbCdsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 554,
label = "Cs-(Cds-O2d)CbCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 555,
label = "Cs-(Cds-Cd)CbCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 556,
label = "Cs-(Cds-Cds)CbCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,6.7,8.16,8.92,9.34,9.16,7.14],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (2.99,'kcal/mol','+|-',0.26),
S298 = (-34.8,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCdCsCs BOZZELLI =3D Cs/Cs3/Cb + (Cs/Cs3/Cd - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 557,
label = "Cs-(Cds-Cdd)CbCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 558,
label = "Cs-(Cds-Cdd-O2d)CbCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CsCsCd(CCO)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 559,
label = "Cs-(Cds-Cdd-Cd)CbCsCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 560,
label = "Cs-CtCtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.57,5.98,7.51,8.37,9,9.02,8.34],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (1.16,'kcal/mol','+|-',0.26),
S298 = (-35.26,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CtCtCsCs BOZZELLI =3D Cs/Cs3/Ct + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 1835,
label = "Cs-(CtN3t)(CtN3t)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S} {6,T}
3 Ct u0 {1,S} {7,T}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 N3t u0 {2,T}
7 N3t u0 {3,T}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (28.4,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 561,
label = "Cs-CbCtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.57,5.98,7.51,8.37,9,9.02,8.34],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (1.16,'kcal/mol','+|-',0.26),
S298 = (-35.26,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCtCsCs BOZZELLI =3D Cs/Cs3/Cb + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 562,
label = "Cs-CbCbCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.57,5.98,7.51,8.37,9,9.02,8.34],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (1.16,'kcal/mol','+|-',0.26),
S298 = (-35.26,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCbCsCs BENSON""",
longDesc =
u"""
""",
)
entry(
index = 563,
label = "Cs-CdsCdsCdsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 564,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-CsCsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 565,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cs u0 {1,S}
6 C u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([35.99,39.53,39.94,39.09,36.71,34.8,32.51],'J/(mol*K)','+|-',[5.08,5.08,5.08,5.08,5.08,5.08,5.08]),
H298 = (19.9,'kJ/mol','+|-',4.33),
S298 = (-150.69,'J/(mol*K)','+|-',5.92),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 566,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cs u0 {1,S}
6 Cd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 567,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cs u0 {1,S}
6 Cdd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 568,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 569,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 570,
label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 571,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cs u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 572,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 573,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 574,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 575,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 576,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 577,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 578,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 579,
label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cs u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 580,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.32,5.86,7.57,8.54,9.22,9.36,8.45],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (2.54,'kcal/mol','+|-',0.26),
S298 = (-33.96,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CdCdCdCs BOZZELLI =3D Cs/Cs2/Cd2 + (Cs/Cs3/Cd - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 581,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 582,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 583,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 584,
label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 585,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 586,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 587,
label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 Cd u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 588,
label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 589,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cs u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 590,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cs u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 591,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cs u0 {1,S}
9 O2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cs u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cs u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cs u0 {1,S}
9 S2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 592,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cs u0 {1,S}
9 C u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 593,
label = "Cs-CtCdsCdsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 594,
label = "Cs-(Cds-O2d)(Cds-O2d)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 595,
label = "Cs-(Cds-O2d)(Cds-Cd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 596,
label = "Cs-(Cds-O2d)(Cds-Cds)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 597,
label = "Cs-(Cds-O2d)(Cds-Cdd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 598,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 599,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 600,
label = "Cs-(Cds-Cd)(Cds-Cd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 601,
label = "Cs-(Cds-Cds)(Cds-Cds)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 602,
label = "Cs-(Cds-Cdd)(Cds-Cds)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 603,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 604,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 605,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 606,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 607,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 608,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Cs u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 609,
label = "Cs-CbCdsCdsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 610,
label = "Cs-(Cds-O2d)(Cds-O2d)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 611,
label = "Cs-(Cds-O2d)(Cds-Cd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 612,
label = "Cs-(Cds-O2d)(Cds-Cds)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 613,
label = "Cs-(Cds-O2d)(Cds-Cdd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 614,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 615,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 616,
label = "Cs-(Cds-Cd)(Cds-Cd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 617,
label = "Cs-(Cds-Cds)(Cds-Cds)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 618,
label = "Cs-(Cds-Cdd)(Cds-Cds)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 619,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 620,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 621,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 622,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 623,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cs u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 624,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cs u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 625,
label = "Cs-CtCtCdsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 626,
label = "Cs-(Cds-O2d)CtCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 627,
label = "Cs-(Cds-Cd)CtCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CtCtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 628,
label = "Cs-(Cds-Cds)CtCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,7.36,8.89,9.58,9.76,9.16,7.25],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (5.1,'kcal/mol','+|-',0.26),
S298 = (-34.88,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CtCtCdCs BOZZELLI =3D Cs/Cd2/Cs2 + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 629,
label = "Cs-(Cds-Cdd)CtCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtCtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 630,
label = "Cs-(Cds-Cdd-O2d)CtCtCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtCtCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 631,
label = "Cs-(Cds-Cdd-Cd)CtCtCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtCtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 632,
label = "Cs-CbCtCdsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 633,
label = "Cs-(Cds-O2d)CbCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 634,
label = "Cs-(Cds-Cd)CbCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 635,
label = "Cs-(Cds-Cds)CbCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,7.36,8.89,9.58,9.76,9.16,7.25],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (5.1,'kcal/mol','+|-',0.26),
S298 = (-34.88,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCtCdCs BOZZELLI =3D Cs/Cb/Cd/Cs2 + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 636,
label = "Cs-(Cds-Cdd)CbCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 637,
label = "Cs-(Cds-Cdd-O2d)CbCtCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCtCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 638,
label = "Cs-(Cds-Cdd-Cd)CbCtCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 C u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,7.36,8.89,9.58,9.76,9.16,7.25],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (5.1,'kcal/mol','+|-',0.26),
S298 = (-34.88,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCtCdCs BOZZELLI =3D Cs/Cb/Cd/Cs2 + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 639,
label = "Cs-CbCbCdsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 640,
label = "Cs-(Cds-O2d)CbCbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 641,
label = "Cs-(Cds-Cd)CbCbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 642,
label = "Cs-(Cds-Cds)CbCbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.99,7.36,8.89,9.58,9.76,9.16,7.25],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (5.1,'kcal/mol','+|-',0.26),
S298 = (-34.88,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCbCdCs BOZZELLI =3D Cs/Cs2/Cb2 + (Cs/Cs3/Cd - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 643,
label = "Cs-(Cds-Cdd)CbCbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 644,
label = "Cs-(Cds-Cdd-O2d)CbCbCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCbCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 645,
label = "Cs-(Cds-Cdd-Cd)CbCbCs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCbCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 646,
label = "Cs-CtCtCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.37,8.11,9.55,10.1,10.03,9.36,6.65],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (6.23,'kcal/mol','+|-',0.26),
S298 = (-35.34,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CtCtCtCs BOZZELLI =3D Cs/Cs2/Ct2 + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 647,
label = "Cs-CbCtCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.37,8.11,9.55,10.1,10.03,9.36,6.65],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (6.23,'kcal/mol','+|-',0.26),
S298 = (-35.34,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCtCtCs BOZZELLI =3D Cs/Cs2/Cb/Ct + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 648,
label = "Cs-CbCbCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.37,8.11,9.55,10.1,10.03,9.36,6.65],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (6.43,'kcal/mol','+|-',0.26),
S298 = (-35.34,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCbCtCs BOZZELLI =3D Cs/Cs2/Cb2 + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 649,
label = "Cs-CbCbCbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.37,8.11,9.55,10.1,10.03,9.36,6.65],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (6.23,'kcal/mol','+|-',0.26),
S298 = (-35.34,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCbCbCs BOZZELLI =3D Cs/Cs2/Cb2 + (Cs/Cs3/Cb - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 650,
label = "Cs-CdsCdsCdsCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 651,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-CsCsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 652,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 CO u0 {1,S} {9,D}
5 Cd u0 {1,S} {6,D}
6 C u0 {5,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
9 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 653,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 CO u0 {1,S} {9,D}
5 Cd u0 {1,S} {6,D}
6 Cd u0 {5,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
9 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 654,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 CO u0 {1,S} {9,D}
5 Cd u0 {1,S} {6,D}
6 Cdd u0 {5,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
9 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 655,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 656,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 657,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 CO u0 {1,S} {9,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {1,S} {7,D}
6 C u0 {4,D}
7 C u0 {5,D}
8 O2d u0 {2,D}
9 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([42.49,50.96,52.27,50.54,45.33,41.1,35.7],'J/(mol*K)','+|-',[5.08,5.08,5.08,5.08,5.08,5.08,5.08]),
H298 = (25.2,'kJ/mol','+|-',4.33),
S298 = (-168.67,'J/(mol*K)','+|-',5.92),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 658,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 CO u0 {1,S} {9,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {1,S} {7,D}
6 Cd u0 {4,D}
7 Cd u0 {5,D}
8 O2d u0 {2,D}
9 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 659,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 CO u0 {1,S} {9,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {1,S} {7,D}
6 Cdd u0 {4,D}
7 Cd u0 {5,D}
8 O2d u0 {2,D}
9 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 660,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CO u0 {1,S} {8,D}
4 CO u0 {1,S} {9,D}
5 Cd u0 {1,S} {7,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {5,D}
8 O2d u0 {3,D}
9 O2d u0 {4,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 661,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CO u0 {1,S} {8,D}
4 CO u0 {1,S} {9,D}
5 Cd u0 {1,S} {7,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {5,D}
8 O2d u0 {3,D}
9 O2d u0 {4,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 662,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 CO u0 {1,S} {9,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {1,S} {7,D}
6 Cdd u0 {4,D}
7 Cdd u0 {5,D}
8 O2d u0 {2,D}
9 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 663,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 664,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 665,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 666,
label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 C u0 {3,D}
7 C u0 {4,D}
8 C u0 {5,D}
9 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 667,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 Cd u0 {5,D}
9 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 668,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 Cdd u0 {5,D}
9 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 669,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CO u0 {1,S} {9,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {4,D}
8 Cd u0 {5,D}
9 O2d u0 {3,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 670,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CO u0 {1,S} {9,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {4,D}
8 Cd u0 {5,D}
9 O2d u0 {3,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 671,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cd u0 {3,D}
7 Cdd u0 {4,D}
8 Cdd u0 {5,D}
9 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 672,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CO u0 {1,S} {9,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {5,D}
9 O2d u0 {4,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 673,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CO u0 {1,S} {9,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {5,D}
9 O2d u0 {4,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 674,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CO u0 {1,S} {9,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {5,D}
9 O2d u0 {4,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 675,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 Cdd u0 {5,D}
9 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 676,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
12 O2d u0 {8,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 677,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
12 C u0 {8,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 678,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 679,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 CO u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 680,
label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 C u0 {2,D}
7 C u0 {3,D}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 681,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
""",
thermo = u'Cs-CsCsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 682,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 Cdd u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 683,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 684,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 685,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cdd u0 {4,D}
9 Cdd u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 686,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 687,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 688,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {4,D}
9 Cd u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 689,
label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
9 Cdd u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 690,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cd u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
12 O2d u0 {8,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 691,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cd u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
12 C u0 {8,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 692,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cd u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cd u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
12 S2d u0 {8,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cd u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
12 C u0 {8,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cd u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 693,
label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cd u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 694,
label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
9 Cdd u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 695,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
12 O2d u0 {8,D}
13 O2d u0 {9,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 696,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
12 O2d u0 {8,D}
13 C u0 {9,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 697,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
12 C u0 {8,D}
13 C u0 {9,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 698,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
13 C u0 {9,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
12 S2d u0 {8,D}
13 S2d u0 {9,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
12 S2d u0 {8,D}
13 C u0 {9,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
12 C u0 {8,D}
13 C u0 {9,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
13 C u0 {9,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 699,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cd u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 Cdd u0 {5,D} {13,D}
10 C u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
13 C u0 {9,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 700,
label = "Cs-CtCdsCdsCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 701,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Ct u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 702,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Ct u0 {1,S}
6 C u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 703,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Ct u0 {1,S}
6 Cd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 704,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Ct u0 {1,S}
6 Cdd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 705,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 706,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 707,
label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 708,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Ct u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 709,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Ct u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 710,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 711,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 712,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Ct u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 713,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 714,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 715,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 716,
label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Ct u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 717,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Ct u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 718,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Ct u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 719,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 720,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 721,
label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Ct u0 {1,S}
6 Cd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 722,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 723,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 724,
label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 Cd u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 725,
label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Ct u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 726,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Ct u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 727,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Ct u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 728,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Ct u0 {1,S}
9 O2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Ct u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Ct u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Ct u0 {1,S}
9 S2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 729,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Ct u0 {1,S}
9 C u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 730,
label = "Cs-CbCdsCdsCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 731,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cb u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 732,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cb u0 {1,S}
6 C u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 733,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cb u0 {1,S}
6 Cd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 734,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cb u0 {1,S}
6 Cdd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 735,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 736,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 737,
label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cb u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 738,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cb u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 739,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cb u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 740,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 741,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 742,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cb u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 743,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 744,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 745,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 746,
label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cb u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 747,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cb u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 748,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cb u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 749,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 750,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 751,
label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cb u0 {1,S}
6 Cd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 752,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 753,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 754,
label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 Cd u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 755,
label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cb u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 756,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cb u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 757,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cb u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 758,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cb u0 {1,S}
9 O2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cb u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cb u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cb u0 {1,S}
9 S2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 759,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 Cb u0 {1,S}
9 C u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 760,
label = "Cs-CtCtCdsCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 761,
label = "Cs-(Cds-O2d)(Cds-O2d)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 762,
label = "Cs-(Cds-O2d)(Cds-Cd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 763,
label = "Cs-(Cds-O2d)(Cds-Cds)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 764,
label = "Cs-(Cds-O2d)(Cds-Cdd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 765,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 766,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 767,
label = "Cs-(Cds-Cd)(Cds-Cd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 768,
label = "Cs-(Cds-Cds)(Cds-Cds)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.61,7.3,8.97,9.69,9.84,9.42,7.36],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (5.48,'kcal/mol','+|-',0.26),
S298 = (-34.5,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CtCtCdCd BOZZELLI =3D Cs/Cs/Cd/Ct2 + (Cs/Cs3/Cd - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 769,
label = "Cs-(Cds-Cdd)(Cds-Cds)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 770,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 771,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 772,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 773,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Ct u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 774,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Ct u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Ct u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Ct u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 775,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 Ct u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 776,
label = "Cs-CbCtCdsCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 777,
label = "Cs-(Cds-O2d)(Cds-O2d)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 778,
label = "Cs-(Cds-O2d)(Cds-Cd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 779,
label = "Cs-(Cds-O2d)(Cds-Cds)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 780,
label = "Cs-(Cds-O2d)(Cds-Cdd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 781,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 782,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 783,
label = "Cs-(Cds-Cd)(Cds-Cd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 784,
label = "Cs-(Cds-Cds)(Cds-Cds)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.61,7.3,8.97,9.69,9.84,9.42,7.36],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (5.48,'kcal/mol','+|-',0.26),
S298 = (-34.5,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCtCdCd BOZZELLI =3D Cs/Cs/Cb/Cd2 + (Cs/Cs3/Ct - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 785,
label = "Cs-(Cds-Cdd)(Cds-Cds)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 786,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 787,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 788,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 789,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Ct u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 790,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Ct u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Ct u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Ct u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 791,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Ct u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 792,
label = "Cs-CbCbCdsCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 793,
label = "Cs-(Cds-O2d)(Cds-O2d)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 794,
label = "Cs-(Cds-O2d)(Cds-Cd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 795,
label = "Cs-(Cds-O2d)(Cds-Cds)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 796,
label = "Cs-(Cds-O2d)(Cds-Cdd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 797,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 798,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 799,
label = "Cs-(Cds-Cd)(Cds-Cd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 800,
label = "Cs-(Cds-Cds)(Cds-Cds)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.61,7.3,8.97,9.69,9.84,9.42,7.36],'cal/(mol*K)','+|-',[0.13,0.13,0.13,0.13,0.13,0.13,0.13]),
H298 = (5.48,'kcal/mol','+|-',0.26),
S298 = (-34.5,'cal/(mol*K)','+|-',0.13),
),
shortDesc = u"""Cs-CbCbCdCd BOZZELLI =3D Cs/Cs/Cb2/Cd + (Cs/Cs3/Cd - Cs/Cs4)""",
longDesc =
u"""
""",
)
entry(
index = 801,
label = "Cs-(Cds-Cdd)(Cds-Cds)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 802,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 803,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 804,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 805,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cb u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 806,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cb u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cb u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cb u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 807,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 Cb u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 808,
label = "Cs-CtCtCtCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 809,
label = "Cs-(Cds-O2d)CtCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CtCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 810,
label = "Cs-(Cds-Cds)CtCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 811,
label = "Cs-(Cds-Cdd)CtCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 812,
label = "Cs-(Cds-Cdd-O2d)CtCtCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtCtCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 813,
label = "Cs-(Cds-Cdd-Cd)CtCtCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 814,
label = "Cs-CbCtCtCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 815,
label = "Cs-(Cds-O2d)CbCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 816,
label = "Cs-(Cds-Cd)CbCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 817,
label = "Cs-(Cds-Cds)CbCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 818,
label = "Cs-(Cds-Cdd)CbCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 819,
label = "Cs-(Cds-Cdd-O2d)CbCtCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCtCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 820,
label = "Cs-(Cds-Cdd-Cd)CbCtCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 821,
label = "Cs-CbCbCtCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 822,
label = "Cs-(Cds-O2d)CbCbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 823,
label = "Cs-(Cds-Cd)CbCbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 824,
label = "Cs-(Cds-Cds)CbCbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 825,
label = "Cs-(Cds-Cdd)CbCbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 826,
label = "Cs-(Cds-Cdd-O2d)CbCbCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCbCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 827,
label = "Cs-(Cds-Cdd-Cd)CbCbCt",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCbCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 828,
label = "Cs-CbCbCbCds",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 [Cd,CO] u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 829,
label = "Cs-(Cds-O2d)CbCbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 830,
label = "Cs-(Cds-Cd)CbCbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 831,
label = "Cs-(Cds-Cds)CbCbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 832,
label = "Cs-(Cds-Cdd)CbCbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 833,
label = "Cs-(Cds-Cdd-O2d)CbCbCb",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCbCb",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 834,
label = "Cs-(Cds-Cdd-Cd)CbCbCb",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCbCb',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 835,
label = "Cs-CtCtCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 836,
label = "Cs-CbCtCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 837,
label = "Cs-CbCbCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtCt',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 838,
label = "Cs-CbCbCbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 839,
label = "Cs-CbCbCbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 C u0 {3,D}
7 C u0 {4,D}
8 C u0 {5,D}
9 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 Cdd u0 {5,D}
9 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CS u0 {1,S} {9,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {4,D}
8 Cd u0 {5,D}
9 S2d u0 {3,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CS u0 {1,S} {9,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {4,D}
8 Cd u0 {5,D}
9 S2d u0 {3,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 Cdd u0 {5,D}
9 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
12 C u0 {8,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
12 S2d u0 {8,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cdd u0 {4,D} {12,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
12 C u0 {8,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 Cd u0 {5,D}
9 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {9,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cd u0 {1,S} {8,D}
6 Cd u0 {3,D}
7 Cdd u0 {4,D}
8 Cdd u0 {5,D}
9 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CS u0 {1,S} {9,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {5,D}
9 S2d u0 {4,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CS u0 {1,S} {9,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {5,D}
9 S2d u0 {4,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CS u0 {1,S} {9,D}
5 Cd u0 {1,S} {8,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 Cd u0 {5,D}
9 S2d u0 {4,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)(Cds-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cs u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cs u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCtCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1179,
label = "Cs-C=SCsCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.14,6.63,7.51,7.98,8.33,8.38,8.24],'cal/(mol*K)'),
H298 = (1.36,'kcal/mol'),
S298 = (-33.92,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=SCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Ct u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=SCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=SC=S",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=SCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cb u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=S(Cds-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 CS u0 {1,S} {9,D}
5 Cd u0 {1,S} {6,D}
6 C u0 {5,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
9 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=S(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 CS u0 {1,S} {9,D}
5 Cd u0 {1,S} {6,D}
6 Cdd u0 {5,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
9 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=S(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=S(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=S(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 CS u0 {1,S} {9,D}
5 Cd u0 {1,S} {6,D}
6 Cd u0 {5,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
9 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)(Cds-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Ct u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Ct u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Ct u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Ct u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CbCt",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cb u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Ct u0 {1,S}
6 C u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cds)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Ct u0 {1,S}
6 Cd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Ct u0 {1,S}
6 Cdd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-Cd)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-S2d)Ct",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Ct u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cs u0 {1,S}
6 C u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cds)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cs u0 {1,S}
6 Cd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cs u0 {1,S}
6 Cdd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-S2d)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-Cd)Cs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cd)(Cds-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 CS u0 {1,S} {9,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {1,S} {7,D}
6 C u0 {4,D}
7 C u0 {5,D}
8 S2d u0 {2,D}
9 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 CS u0 {1,S} {9,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {1,S} {7,D}
6 Cdd u0 {4,D}
7 Cd u0 {5,D}
8 S2d u0 {2,D}
9 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CS u0 {1,S} {8,D}
4 CS u0 {1,S} {9,D}
5 Cd u0 {1,S} {7,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {5,D}
8 S2d u0 {3,D}
9 S2d u0 {4,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-Cd)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 CS u0 {1,S} {8,D}
4 CS u0 {1,S} {9,D}
5 Cd u0 {1,S} {7,D}
6 Cdd u0 {2,D} {10,D}
7 Cd u0 {5,D}
8 S2d u0 {3,D}
9 S2d u0 {4,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd)(Cds-Cdd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 CS u0 {1,S} {9,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {1,S} {7,D}
6 Cdd u0 {4,D}
7 Cdd u0 {5,D}
8 S2d u0 {2,D}
9 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 CS u0 {1,S} {9,D}
6 Cdd u0 {2,D} {10,D}
7 Cdd u0 {3,D} {11,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cds)(Cds-Cds)",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 CS u0 {1,S} {9,D}
4 Cd u0 {1,S} {6,D}
5 Cd u0 {1,S} {7,D}
6 Cd u0 {4,D}
7 Cd u0 {5,D}
8 S2d u0 {2,D}
9 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cb u0 {1,S}
6 C u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cb u0 {1,S}
6 Cdd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-S2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 Cb u0 {1,S}
6 Cd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCtCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 Cs u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)(Cds-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cb u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cb u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cb u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cb u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cb u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Cb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cb u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCbCb",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCbCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 840,
label = "Cs-CCCOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 C u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-CsCsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 841,
label = "Cs-CsCsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([23.99,31.2,34.89,36.47,36.78,36.05,34.4],'J/(mol*K)','+|-',[3.81,3.81,3.81,3.81,3.81,3.81,3.81]),
H298 = (-20.3,'kJ/mol','+|-',3.24),
S298 = (-144.38,'J/(mol*K)','+|-',4.44),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 842,
label = "Cs-CdsCsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 843,
label = "Cs-(Cds-O2d)CsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([28.15,35.17,38.11,38.72,37.49,35.88,33.45],'J/(mol*K)','+|-',[5.16,5.16,5.16,5.16,5.16,5.16,5.16]),
H298 = (-10.9,'kJ/mol','+|-',4.39),
S298 = (-148.7,'J/(mol*K)','+|-',6.02),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 844,
label = "Cs-(Cds-Cd)CsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([29.24,37.61,40.84,41.46,40.06,38.2,35.08],'J/(mol*K)','+|-',[3.81,3.81,3.81,3.81,3.81,3.81,3.81]),
H298 = (-14.6,'kJ/mol','+|-',3.24),
S298 = (-153.23,'J/(mol*K)','+|-',4.44),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 845,
label = "Cs-(Cds-Cds)CsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.63,6.79,7.95,8.4,8.8,8.44,8.44],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]),
H298 = (-6.6,'kcal/mol','+|-',0.4),
S298 = (-32.56,'cal/(mol*K)','+|-',0.2),
),
shortDesc = u"""Cs-OCdCsCs BOZZELLI C/C3/O - (C/C3/H - C/Cb/C2/H), Hf-1 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 846,
label = "Cs-(Cds-Cdd)CsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 847,
label = "Cs-(Cds-Cdd-O2d)CsCsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.39,9.66,10.03,10.07,9.64,9.26,8.74],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]),
H298 = (-9.725,'kcal/mol','+|-',0.4),
S298 = (-36.5,'cal/(mol*K)','+|-',0.2),
),
shortDesc = u"""{C/CCO/O/C2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 848,
label = "Cs-(Cds-Cdd-Cd)CsCsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 849,
label = "Cs-OsCtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 O2s u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 850,
label = "Cs-CbCsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.63,6.79,7.95,8.4,8.8,8.44,8.44],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]),
H298 = (-6.6,'kcal/mol','+|-',0.4),
S298 = (-32.56,'cal/(mol*K)','+|-',0.2),
),
shortDesc = u"""Cs-OCbCsCs BOZZELLI C/C3/O - (C/C3/H - C/Cb/C2/H), Hf-1 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 851,
label = "Cs-CdsCdsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 852,
label = "Cs-(Cds-O2d)(Cds-O2d)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CsCsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 853,
label = "Cs-(Cds-O2d)(Cds-Cd)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([33.75,42.15,45.09,44.95,41.74,38.55,34.46],'J/(mol*K)','+|-',[4.3,4.3,4.3,4.3,4.3,4.3,4.3]),
H298 = (-3.9,'kJ/mol','+|-',3.66),
S298 = (-158.3,'J/(mol*K)','+|-',5.02),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 854,
label = "Cs-(Cds-O2d)(Cds-Cds)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 855,
label = "Cs-(Cds-O2d)(Cds-Cdd)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 856,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 857,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 858,
label = "Cs-(Cds-Cd)(Cds-Cd)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 859,
label = "Cs-(Cds-Cds)(Cds-Cds)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.61,5.98,7.51,8.37,9,9.02,8.34],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]),
H298 = (-8.01,'kcal/mol','+|-',0.4),
S298 = (-34.34,'cal/(mol*K)','+|-',0.2),
),
shortDesc = u"""Cs-OCdCdCs Hf jwb 697 S,Cp from C/Cd2/C2""",
longDesc =
u"""
""",
)
entry(
index = 860,
label = "Cs-(Cds-Cdd)(Cds-Cds)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 861,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 862,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 863,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 864,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 865,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 866,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 O2s u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 867,
label = "Cs-CtCdsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 868,
label = "Cs-(Cds-O2d)CtCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 869,
label = "Cs-(Cds-Cd)CtCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CtCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 870,
label = "Cs-(Cds-Cds)CtCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 871,
label = "Cs-(Cds-Cdd)CtCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 872,
label = "Cs-(Cds-Cdd-O2d)CtCsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 873,
label = "Cs-(Cds-Cdd-Cd)CtCsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 874,
label = "Cs-CbCdsCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 875,
label = "Cs-(Cds-O2d)CbCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 876,
label = "Cs-(Cds-Cd)CbCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 877,
label = "Cs-(Cds-Cds)CbCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 878,
label = "Cs-(Cds-Cdd)CbCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 879,
label = "Cs-(Cds-Cdd-O2d)CbCsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 880,
label = "Cs-(Cds-Cdd-Cd)CbCsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 881,
label = "Cs-CtCtCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 882,
label = "Cs-CbCtCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 883,
label = "Cs-CbCbCsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 884,
label = "Cs-CdsCdsCdsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 885,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-CsCsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 886,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 O2s u0 {1,S}
6 C u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 887,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 O2s u0 {1,S}
6 Cd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 888,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 O2s u0 {1,S}
6 Cdd u0 {4,D}
7 O2d u0 {2,D}
8 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 889,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 890,
label = "Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {7,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 891,
label = "Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 O2s u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([36.85,46.04,49,48.85,45.61,42.23,37.25],'J/(mol*K)','+|-',[4.09,4.09,4.09,4.09,4.09,4.09,4.09]),
H298 = (3,'kJ/mol','+|-',3.49),
S298 = (-160.69,'J/(mol*K)','+|-',4.77),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 892,
label = "Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 O2s u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 893,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 O2s u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 894,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 O2s u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 895,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CO u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 O2s u0 {1,S}
7 Cd u0 {4,D}
8 O2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 896,
label = "Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 O2s u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 897,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 898,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 899,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CO u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 900,
label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 O2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 901,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
""",
thermo = u'Cs-CsCsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 902,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 903,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsCsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 904,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 905,
label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 906,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 O2s u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 907,
label = "Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 O2s u0 {1,S}
8 Cd u0 {4,D}
9 O2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 908,
label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 O2s u0 {1,S}
8 Cd u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 909,
label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 910,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 O2s u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 O2d u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 911,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 O2s u0 {1,S}
9 O2d u0 {5,D}
10 O2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 912,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 O2s u0 {1,S}
9 O2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 913,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 O2s u0 {1,S}
9 C u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 914,
label = "Cs-CtCdsCdsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 915,
label = "Cs-(Cds-O2d)(Cds-O2d)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 916,
label = "Cs-(Cds-O2d)(Cds-Cd)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 917,
label = "Cs-(Cds-O2d)(Cds-Cds)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 918,
label = "Cs-(Cds-O2d)(Cds-Cdd)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 919,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 920,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 921,
label = "Cs-(Cds-Cd)(Cds-Cd)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 922,
label = "Cs-(Cds-Cds)(Cds-Cds)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 923,
label = "Cs-(Cds-Cdd)(Cds-Cds)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 924,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 925,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 926,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 927,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 928,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 929,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 O2s u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 930,
label = "Cs-CbCdsCdsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 931,
label = "Cs-(Cds-O2d)(Cds-O2d)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 932,
label = "Cs-(Cds-O2d)(Cds-Cd)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 933,
label = "Cs-(Cds-O2d)(Cds-Cds)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 934,
label = "Cs-(Cds-O2d)(Cds-Cdd)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 935,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 936,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 937,
label = "Cs-(Cds-Cd)(Cds-Cd)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 938,
label = "Cs-(Cds-Cds)(Cds-Cds)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 939,
label = "Cs-(Cds-Cdd)(Cds-Cds)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 940,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 941,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 942,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 943,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 944,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 945,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 O2s u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 946,
label = "Cs-CtCtCdsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 947,
label = "Cs-(Cds-O2d)CtCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 948,
label = "Cs-(Cds-Cd)CtCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CtCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 949,
label = "Cs-(Cds-Cds)CtCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 950,
label = "Cs-(Cds-Cdd)CtCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 951,
label = "Cs-(Cds-Cdd-O2d)CtCtOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 952,
label = "Cs-(Cds-Cdd-Cd)CtCtOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 953,
label = "Cs-CbCtCdsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 954,
label = "Cs-(Cds-O2d)CbCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 955,
label = "Cs-(Cds-Cd)CbCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 956,
label = "Cs-(Cds-Cds)CbCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 957,
label = "Cs-(Cds-Cdd)CbCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 958,
label = "Cs-(Cds-Cdd-O2d)CbCtOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 959,
label = "Cs-(Cds-Cdd-Cd)CbCtOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 960,
label = "Cs-CbCbCdsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 [Cd,CO] u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbCbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 961,
label = "Cs-(Cds-O2d)CbCbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 962,
label = "Cs-(Cds-Cd)CbCbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbCbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 963,
label = "Cs-(Cds-Cds)CbCbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 964,
label = "Cs-(Cds-Cdd)CbCbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbCbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 965,
label = "Cs-(Cds-Cdd-O2d)CbCbOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 966,
label = "Cs-(Cds-Cdd-Cd)CbCbOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbCbOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 967,
label = "Cs-CtCtCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 968,
label = "Cs-CbCtCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtCtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 969,
label = "Cs-CbCbCtOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)CtOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 970,
label = "Cs-CbCbCbOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 971,
label = "Cs-CCOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-CsCsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 972,
label = "Cs-CsCsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([27.88,43.75,51.85,54,50.77,45.94,38.31],'J/(mol*K)','+|-',[5.77,5.77,5.77,5.77,5.77,5.77,5.77]),
H298 = (-69.2,'kJ/mol','+|-',4.92),
S298 = (-163.77,'J/(mol*K)','+|-',6.74),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 973,
label = "Cs-CdsCsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 974,
label = "Cs-(Cds-O2d)CsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-CsCsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 975,
label = "Cs-(Cds-Cd)CsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([27.95,42.92,51.33,54.81,53.92,49.73,41.11],'J/(mol*K)','+|-',[5.77,5.77,5.77,5.77,5.77,5.77,5.77]),
H298 = (-62.8,'kJ/mol','+|-',4.92),
S298 = (-170.44,'J/(mol*K)','+|-',6.74),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 976,
label = "Cs-(Cds-Cds)CsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-CsCsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 977,
label = "Cs-(Cds-Cdd)CsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 978,
label = "Cs-(Cds-Cdd-O2d)CsOsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 979,
label = "Cs-(Cds-Cdd-Cd)CsOsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 980,
label = "Cs-CdsCdsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 981,
label = "Cs-(Cds-O2d)(Cds-O2d)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CsCsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 982,
label = "Cs-(Cds-O2d)(Cds-Cd)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 983,
label = "Cs-(Cds-O2d)(Cds-Cds)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 984,
label = "Cs-(Cds-O2d)(Cds-Cdd)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 985,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 986,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 987,
label = "Cs-(Cds-Cd)(Cds-Cd)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([30.08,45.85,54.7,58.39,57.78,53.65,44.31],'J/(mol*K)','+|-',[5.77,5.77,5.77,5.77,5.77,5.77,5.77]),
H298 = (-55.7,'kJ/mol','+|-',4.92),
S298 = (-179.76,'J/(mol*K)','+|-',6.74),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 988,
label = "Cs-(Cds-Cds)(Cds-Cds)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-CsCsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 989,
label = "Cs-(Cds-Cdd)(Cds-Cds)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 990,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 991,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 992,
label = "Cs-(Cds-Cdd)(Cds-Cdd)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 993,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 O2s u0 {1,S}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 994,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 O2s u0 {1,S}
7 O2s u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 995,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 O2s u0 {1,S}
7 O2s u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 996,
label = "Cs-CtCsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 997,
label = "Cs-CtCdsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 998,
label = "Cs-(Cds-O2d)CtOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 999,
label = "Cs-(Cds-Cd)CtOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CtOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1000,
label = "Cs-(Cds-Cds)CtOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1001,
label = "Cs-(Cds-Cdd)CtOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1002,
label = "Cs-(Cds-Cdd-O2d)CtOsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1003,
label = "Cs-(Cds-Cdd-Cd)CtOsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1004,
label = "Cs-CtCtOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1005,
label = "Cs-CbCsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1006,
label = "Cs-CbCdsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1007,
label = "Cs-(Cds-O2d)CbOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1008,
label = "Cs-(Cds-Cd)CbOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1009,
label = "Cs-(Cds-Cds)CbOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1010,
label = "Cs-(Cds-Cdd)CbOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1011,
label = "Cs-(Cds-Cdd-O2d)CbOsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1012,
label = "Cs-(Cds-Cdd-Cd)CbOsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1013,
label = "Cs-CbCtOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1014,
label = "Cs-CbCbOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1015,
label = "Cs-COsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-CsOsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1016,
label = "Cs-CsOsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.33,6.19,7.25,7.7,8.2,8.24,8.24],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]),
H298 = (-19,'kcal/mol','+|-',0.4),
S298 = (-33.56,'cal/(mol*K)','+|-',0.2),
),
shortDesc = u"""Cs-OOOCs BOZZELLI est !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1017,
label = "Cs-CdsOsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)OsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1018,
label = "Cs-(Cds-O2d)OsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-CsOsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1019,
label = "Cs-(Cds-Cd)OsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)OsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1020,
label = "Cs-(Cds-Cds)OsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-CsOsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1021,
label = "Cs-(Cds-Cdd)OsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)OsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1022,
label = "Cs-(Cds-Cdd-O2d)OsOsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)OsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1023,
label = "Cs-(Cds-Cdd-Cd)OsOsOs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 O2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)OsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1024,
label = "Cs-CtOsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)OsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1025,
label = "Cs-CbOsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)OsOsOs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1026,
label = "Cs-OsOsOsOs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 O2s u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.33,6.13,7.25,7.7,8.2,8.24,8.24],'cal/(mol*K)','+|-',[0.2,0.2,0.2,0.2,0.2,0.2,0.2]),
H298 = (-23,'kcal/mol','+|-',0.4),
S298 = (-35.56,'cal/(mol*K)','+|-',0.2),
),
shortDesc = u"""Cs-OOOO BOZZELLI est !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1027,
label = "Cs-COsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CsOsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1028,
label = "Cs-CsOsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.25,7.1,8.81,9.55,10.31,11.05,11.05],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]),
H298 = (-16,'kcal/mol','+|-',0.24),
S298 = (-12.07,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cs-OOCsH BENSON Hf, BOZZELLI C/C3/H - C/C2/O/H !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1029,
label = "Cs-CdsOsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)OsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1030,
label = "Cs-(Cds-O2d)OsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-CsOsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1031,
label = "Cs-(Cds-Cd)OsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)OsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1032,
label = "Cs-(Cds-Cds)OsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-CsOsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1033,
label = "Cs-(Cds-Cdd)OsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)OsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1034,
label = "Cs-(Cds-Cdd-O2d)OsOsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)OsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1035,
label = "Cs-(Cds-Cdd-Cd)OsOsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 O2s u0 {1,S}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)OsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1036,
label = "Cs-CtOsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)OsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1037,
label = "Cs-CbOsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)OsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-COsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 O2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1452,
label = "Cs-CsOsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 O2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.37,10.32,11.1,11.3,11.3,11.21,11.6],'cal/(mol*K)'),
H298 = (-11.1,'kcal/mol'),
S298 = (-16.14,'cal/(mol*K)'),
),
shortDesc = u"""CAC CBS-QB3 1DHR calc""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsOsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 O2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtOsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 O2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbOsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 O2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CCOsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 O2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1453,
label = "Cs-CsCsOsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.16,10.15,10.69,10.52,9.74,9.01,8.34],'cal/(mol*K)'),
H298 = (-11.26,'kcal/mol'),
S298 = (-39.73,'cal/(mol*K)'),
),
shortDesc = u"""CAC CBS-QB3 1DHR calc""",
longDesc =
u"""
""",
)
entry(
index = 1467,
label = "Cs-COsOsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = u'Cs-CsOsOsSs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1465,
label = "Cs-CsOsOsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.65,8.43,9.23,9.47,9.43,9.2,8.89],'cal/(mol*K)'),
H298 = (-21.41,'kcal/mol'),
S298 = (-36.7,'cal/(mol*K)'),
),
shortDesc = u"""CAC calc 1D-HR""",
longDesc =
u"""
""",
)
entry(
index = 1038,
label = "Cs-CCOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CsCsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1039,
label = "Cs-CsCsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([21.99,29.03,34.22,37.78,41.96,44.27,47.11],'J/(mol*K)','+|-',[3.32,3.32,3.32,3.32,3.32,3.32,3.32]),
H298 = (-25.1,'kJ/mol','+|-',2.83),
S298 = (-52.05,'J/(mol*K)','+|-',3.88),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1040,
label = "Cs-CdsCsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1041,
label = "Cs-(Cds-O2d)CsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.47,6.82,8.45,9.17,10.24,10.8,11.02],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]),
H298 = (-6,'kcal/mol','+|-',0.24),
S298 = (-11.1,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cs-OCOCsH BOZZELLI""",
longDesc =
u"""
""",
)
entry(
index = 1042,
label = "Cs-(Cds-Cd)CsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([29.84,38.86,43.83,46.37,48.34,49.06,49.94],'J/(mol*K)','+|-',[3.74,3.74,3.74,3.74,3.74,3.74,3.74]),
H298 = (-24,'kJ/mol','+|-',3.19),
S298 = (-61.06,'J/(mol*K)','+|-',4.36),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1043,
label = "Cs-(Cds-Cds)CsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.47,6.82,8.45,9.17,10.24,10.8,11.02],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]),
H298 = (-6,'kcal/mol','+|-',0.24),
S298 = (-11.1,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cs-OCdCsH BOZZELLI""",
longDesc =
u"""
""",
)
entry(
index = 1044,
label = "Cs-(Cds-Cdd)CsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1045,
label = "Cs-(Cds-Cdd-O2d)CsOsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.2,8.49,9.33,9.92,10.5,10.92,11.71],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]),
H298 = (-8.37,'kcal/mol','+|-',0.24),
S298 = (-13.04,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""{C/CCO/O/C/H} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 1046,
label = "Cs-(Cds-Cdd-Cd)CsOsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1047,
label = "Cs-CdsCdsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1048,
label = "Cs-(Cds-O2d)(Cds-O2d)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 CO u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
7 O2d u0 {3,D}
""",
thermo = u'Cs-CsCsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1049,
label = "Cs-(Cds-O2d)(Cds-Cd)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1050,
label = "Cs-(Cds-O2d)(Cds-Cds)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)CsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1051,
label = "Cs-(Cds-O2d)(Cds-Cdd)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cdd-Cd)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1052,
label = "Cs-(Cds-O2d)(Cds-Cdd-O2d)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1053,
label = "Cs-(Cds-O2d)(Cds-Cdd-Cd)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CO u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1054,
label = "Cs-(Cds-Cd)(Cds-Cd)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([29.82,38.47,43.27,45.7,47.5,48.09,48.78],'J/(mol*K)','+|-',[3.64,3.64,3.64,3.64,3.64,3.64,3.64]),
H298 = (-17.4,'kJ/mol','+|-',3.1),
S298 = (-64.14,'J/(mol*K)','+|-',4.24),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1055,
label = "Cs-(Cds-Cds)(Cds-Cds)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.21,6.6,8.26,9.05,10.23,10.86,11.04],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]),
H298 = (-6.67,'kcal/mol','+|-',0.24),
S298 = (-10.42,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cs-OCdCdH BOZZELLI""",
longDesc =
u"""
""",
)
entry(
index = 1056,
label = "Cs-(Cds-Cdd)(Cds-Cds)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1057,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 O2d u0 {4,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)CsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1058,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1059,
label = "Cs-(Cds-Cdd)(Cds-Cdd)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1060,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 O2s u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 O2d u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1061,
label = "Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 O2s u0 {1,S}
7 H u0 {1,S}
8 O2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1062,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 O2s u0 {1,S}
7 H u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1063,
label = "Cs-CtCsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CsOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1064,
label = "Cs-CtCdsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1065,
label = "Cs-(Cds-O2d)CtOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1066,
label = "Cs-(Cds-Cd)CtOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CtOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1067,
label = "Cs-(Cds-Cds)CtOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1068,
label = "Cs-(Cds-Cdd)CtOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CtOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1069,
label = "Cs-(Cds-Cdd-O2d)CtOsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1070,
label = "Cs-(Cds-Cdd-Cd)CtOsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CtOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1071,
label = "Cs-CtCtOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1072,
label = "Cs-CbCsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.47,6.82,8.45,9.17,10.24,10.8,11.02],'cal/(mol*K)','+|-',[0.12,0.12,0.12,0.12,0.12,0.12,0.12]),
H298 = (-6,'kcal/mol','+|-',0.24),
S298 = (-11.1,'cal/(mol*K)','+|-',0.12),
),
shortDesc = u"""Cs-OCbCsH BOZZELLI =3D C/Cd/C/H/O Jul 91""",
longDesc =
u"""
""",
)
entry(
index = 1073,
label = "Cs-CbCdsOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 [Cd,CO] u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CbOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1074,
label = "Cs-(Cds-O2d)CbOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = u'Cs-(Cds-O2d)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1075,
label = "Cs-(Cds-Cd)CbOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)CbOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1076,
label = "Cs-(Cds-Cds)CbOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1077,
label = "Cs-(Cds-Cdd)CbOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)CbOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1078,
label = "Cs-(Cds-Cdd-O2d)CbOsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = u'Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1079,
label = "Cs-(Cds-Cdd-Cd)CbOsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 O2s u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)CbOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1080,
label = "Cs-CbCtOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)CtOsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1081,
label = "Cs-CbCbOsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 O2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)(Cds-Cds)OsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1082,
label = "Cs-COsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-CsOsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1083,
label = "Cs-CsOsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([25.01,31.9,37.45,41.88,48.53,53.31,60.53],'J/(mol*K)','+|-',[1.43,1.43,1.43,1.43,1.43,1.43,1.43]),
H298 = (-34.3,'kJ/mol','+|-',1.22),
S298 = (37.65,'J/(mol*K)','+|-',1.67),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1084,
label = "Cs-CdsOsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 [Cd,CO] u0 {1,S}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)OsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1085,
label = "Cs-(Cds-O2d)OsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CO u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([26.75,34.37,40.77,45.37,51.2,54.96,60.79],'J/(mol*K)','+|-',[4.34,4.34,4.34,4.34,4.34,4.34,4.34]),
H298 = (-19.8,'kJ/mol','+|-',3.7),
S298 = (31.54,'J/(mol*K)','+|-',5.06),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1086,
label = "Cs-(Cds-Cd)OsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([28.42,35.65,40.62,44.31,49.79,53.92,60.6],'J/(mol*K)','+|-',[3.38,3.38,3.38,3.38,3.38,3.38,3.38]),
H298 = (-26.6,'kJ/mol','+|-',2.88),
S298 = (34.59,'J/(mol*K)','+|-',3.95),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1087,
label = "Cs-(Cds-Cds)OsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.12,6.86,8.32,9.49,11.22,12.48,14.4],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-6.76,'kcal/mol','+|-',0.2),
S298 = (9.8,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cs-OCdHH BOZZELLI Hf PEDLEY c*ccoh C/C/Cd/H2""",
longDesc =
u"""
""",
)
entry(
index = 1088,
label = "Cs-(Cds-Cdd)OsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = u'Cs-(Cds-Cdd-Cd)OsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1089,
label = "Cs-(Cds-Cdd-O2d)OsHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.15,8.67,9.75,10.65,11.93,12.97,14.86],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-8.68,'kcal/mol','+|-',0.2),
S298 = (8.43,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""{C/CCO/O/H2} RAMAN & GREEN JPCA 2002, 106, 7937-7949""",
longDesc =
u"""
""",
)
entry(
index = 1090,
label = "Cs-(Cds-Cdd-Cd)OsHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 O2s u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = u'Cs-(Cds-Cds)OsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1091,
label = "Cs-CtOsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.12,6.86,8.32,9.49,11.22,12.48,14.4],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-6.76,'kcal/mol','+|-',0.2),
S298 = (9.8,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""Cs-OCtHH BOZZELLI assigned C/Cd/H2/O""",
longDesc =
u"""
""",
)
entry(
index = 1092,
label = "Cs-CbOsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 O2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = u'Cs-(Cds-Cds)OsHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CCCSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 C u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1175,
label = "Cs-CsCsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.41,7.05,8.02,8.53,8.87,8.85,8.57],'cal/(mol*K)'),
H298 = (-0.49,'kcal/mol'),
S298 = (-34.44,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsCsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CsCsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CsCsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-SsCtCsCs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 S2s u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsCdsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 Cd u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)(Cds-Cd)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cds)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cs u0 {1,S}
7 S2s u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCdsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cd u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CtCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CtCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CtCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtCsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CtCsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCdsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cd u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CbCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CbCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CbCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CbCsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCtCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCtCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCbCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsCdsCdsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 Cd u0 {1,S}
4 Cd u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 S2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 Cd u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 S2s u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 S2s u0 {1,S}
8 Cd u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 S2s u0 {1,S}
8 Cd u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cd u0 {1,S} {8,D}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
8 Cdd u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 S2s u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 S2d u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 S2s u0 {1,S}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 S2s u0 {1,S}
9 S2d u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {8,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 Cdd u0 {4,D} {11,D}
8 S2s u0 {1,S}
9 C u0 {5,D}
10 C u0 {6,D}
11 C u0 {7,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCdsCdsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cd u0 {1,S}
4 Cd u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)(Cds-Cd)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cds)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Ct u0 {1,S}
7 S2s u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCdsCdsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cd u0 {1,S}
4 Cd u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)(Cds-Cd)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cds)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cdd)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 Cb u0 {1,S}
7 S2s u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCtCdsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Cd u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CtCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CtCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CtCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtCtSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CtCtSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 Ct u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCtCdsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Cd u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CbCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CbCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CbCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCtSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CbCtSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Ct u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCbCdsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cd u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CbCbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CbCbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CbCbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbCbSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CbCbSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 Cb u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCtCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCtCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCbCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCbCbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCsCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)(Cds-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 S2s u0 {1,S}
6 C u0 {3,D}
7 C u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cdd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 S2s u0 {1,S}
6 Cdd u0 {3,D}
7 Cdd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 C u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {7,S}
2 Cd u0 {1,S} {5,D}
3 Cd u0 {1,S} {6,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 Cdd u0 {3,D} {10,D}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
10 S2d u0 {6,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)(Cds-Cds)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 S2s u0 {1,S}
6 Cdd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 S2s u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {7,D}
5 Cdd u0 {2,D} {9,D}
6 S2s u0 {1,S}
7 Cd u0 {4,D}
8 S2d u0 {3,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)(Cds-Cds)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {8,D}
3 Cd u0 {1,S} {6,D}
4 Cd u0 {1,S} {7,D}
5 S2s u0 {1,S}
6 Cd u0 {3,D}
7 Cd u0 {4,D}
8 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Ct u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SC=SSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 S2s u0 {1,S}
6 C u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cds)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 S2s u0 {1,S}
6 Cd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 CS u0 {1,S} {8,D}
4 Cd u0 {1,S} {6,D}
5 S2s u0 {1,S}
6 Cdd u0 {4,D}
7 S2d u0 {2,D}
8 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-S2d)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=S(Cds-Cdd-Cd)S2s",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {6,S}
2 Cd u0 {1,S} {5,D}
3 CS u0 {1,S} {7,D}
4 CS u0 {1,S} {8,D}
5 Cdd u0 {2,D} {9,D}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cb u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CbSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SCtSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)CsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 Cs u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CCSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1202,
label = "Cs-CsCsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.17,8.72,9.4,9.63,9.55,9.29,8.67],'cal/(mol*K)'),
H298 = (-1.34,'kcal/mol'),
S298 = (-36.66,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsCsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CsSsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CsSsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsCdsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)(Cds-Cd)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cds)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cdd)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 S2s u0 {1,S}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 S2s u0 {1,S}
7 S2s u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 S2s u0 {1,S}
7 S2s u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCdsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CtSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CtSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CtSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtSsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CtSsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCtSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCdsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CbSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CbSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CbSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbSsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CbSsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCtSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCbSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)SsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CSsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1203,
label = "Cs-CsSsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.45,9.59,10.49,10.71,10.42,9.94,8.92],'cal/(mol*K)'),
H298 = (-1.8,'kcal/mol'),
S298 = (-38.19,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsSsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)SsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)SsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)SsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)SsSsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)SsSsSs",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 S2s u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtSsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbSsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SSsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-SsSsSsSs",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 S2s u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CSsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1173,
label = "Cs-CsSsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.82,9.05,9.73,10.14,10.63,10.97,11.44],'cal/(mol*K)'),
H298 = (-3.3,'kcal/mol'),
S298 = (-14.59,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsSsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)SsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)SsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)SsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)SsSsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)SsSsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 S2s u0 {1,S}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtSsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbSsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SSsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CCSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1169,
label = "Cs-CsCsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.01,6.7,7.96,8.83,9.89,10.51,11.27],'cal/(mol*K)'),
H298 = (-1.98,'kcal/mol'),
S298 = (-11.89,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1170,
label = "Cs-CdsCsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.41,9.85,10.93,11.28,11.37,11.41,11.61],'cal/(mol*K)'),
H298 = (-2.15,'kcal/mol'),
S298 = (-15.26,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CsSsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CsSsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cs u0 {1,S}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CdsCdsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)(Cds-Cd)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)(Cds-Cds)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cds)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cds)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cds)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 Cd u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)(Cds-Cdd)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cd u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
7 Cdd u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 S2s u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 S2d u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 S2s u0 {1,S}
7 H u0 {1,S}
8 S2d u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {6,S} {7,S}
2 Cd u0 {1,S} {4,D}
3 Cd u0 {1,S} {5,D}
4 Cdd u0 {2,D} {8,D}
5 Cdd u0 {3,D} {9,D}
6 S2s u0 {1,S}
7 H u0 {1,S}
8 C u0 {4,D}
9 C u0 {5,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1171,
label = "Cs-CtCsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.08,6.78,7.93,8.72,9.73,10.35,11.19],'cal/(mol*K)'),
H298 = (0.72,'kcal/mol'),
S298 = (-11.64,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCdsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CtSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CtSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CtSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CtSsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CtSsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Ct u0 {1,S}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CtCtSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1172,
label = "Cs-CbCsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.38,7.96,9.03,9.69,10.45,10.89,11.47],'cal/(mol*K)'),
H298 = (-1.66,'kcal/mol'),
S298 = (-13.65,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCdsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cd u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)CbSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)CbSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)CbSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)CbSsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)CbSsH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 Cb u0 {1,S}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCtSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CbCbSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCbSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cb u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SC=SSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 CS u0 {1,S} {7,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1174,
label = "Cs-C=SCsSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Cs u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([7.53,10.09,11.25,11.65,11.76,11.74,11.77],'cal/(mol*K)'),
H298 = (-3.49,'kcal/mol'),
S298 = (-15.86,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=SCtSsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 Ct u0 {1,S}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cd)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 C u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-Cd)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 C u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cdd-S2d)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {5,S} {6,S}
2 Cd u0 {1,S} {4,D}
3 CS u0 {1,S} {7,D}
4 Cdd u0 {2,D} {8,D}
5 S2s u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
8 S2d u0 {4,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-C=S(Cds-Cds)SsH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {7,D}
3 Cd u0 {1,S} {6,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {3,D}
7 S2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-CSsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 C u0 {1,S}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1163,
label = "Cs-CsSsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cs u0 {1,S}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.11,6.59,7.97,9.15,10.99,12.33,14.32],'cal/(mol*K)'),
H298 = (-4.94,'kcal/mol'),
S298 = (9.92,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1164,
label = "Cs-CdsSsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.17,9.71,10.55,11.14,12.11,12.95,14.43],'cal/(mol*K)'),
H298 = (-5.07,'kcal/mol'),
S298 = (6.75,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cd)SsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cds)SsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd)SsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cd u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 Cdd u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-S2d)SsHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 S2d u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "Cs-(Cds-Cdd-Cd)SsHH",
group =
"""
1 * Cs u0 {2,S} {4,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D}
3 Cdd u0 {2,D} {7,D}
4 S2s u0 {1,S}
5 H u0 {1,S}
6 H u0 {1,S}
7 C u0 {3,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1165,
label = "Cs-CtSsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Ct u0 {1,S}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.19,7.02,8.42,9.51,11.14,12.34,14.18],'cal/(mol*K)'),
H298 = (-2.69,'kcal/mol'),
S298 = (9.75,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1166,
label = "Cs-CbSsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 Cb u0 {1,S}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.3,7.96,9.19,10.12,11.53,12.59,14.26],'cal/(mol*K)'),
H298 = (-5.04,'kcal/mol'),
S298 = (8.26,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1168,
label = "Cs-C=SSsHH",
group =
"""
1 * Cs u0 {2,S} {3,S} {4,S} {5,S}
2 CS u0 {1,S} {6,D}
3 S2s u0 {1,S}
4 H u0 {1,S}
5 H u0 {1,S}
6 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.23,10.43,11.42,11.97,12.76,13.43,14.63],'cal/(mol*K)'),
H298 = (-6.13,'kcal/mol'),
S298 = (5.73,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1093,
label = "O",
group =
"""
1 * O u0
""",
thermo = u'O2s-CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1094,
label = "O2d",
group =
"""
1 * O2d u0
""",
thermo = u'O2d-Cd',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1095,
label = "O2d-Cd",
group =
"""
1 * O2d u0 {2,D}
2 CO u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""In this case the C is treated as the central atom""",
longDesc =
u"""
""",
)
entry(
index = 1096,
label = "O2d-O2d",
group =
"""
1 * O2d u0 {2,D}
2 O2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.5,3.575,3.685,3.8,3.99,4.12,4.29],'cal/(mol*K)'),
H298 = (14.01,'kcal/mol'),
S298 = (24.085,'cal/(mol*K)'),
),
shortDesc = u"""A. Vandeputte""",
longDesc =
u"""
""",
)
entry(
index = 1943,
label = "O2d-N3d",
group =
"""
1 * O2d u0 {2,D}
2 N3d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1944,
label = "O2d-N5dc",
group =
"""
1 * O2d u0 {2,D}
2 N5dc u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1097,
label = "O2s",
group =
"""
1 * O2s u0
""",
thermo = u'O2s-(Cds-Cd)(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1945,
label = "O2s-N",
group =
"""
1 * O2s u0 {2,S}
2 N u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1935,
label = "O2s-CN",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 C u0 {1,S}
3 N u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1874,
label = "O2s-CsN3s",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 N3s u0 {1,S}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.5,3.6,4,4.3,4.7,4.8,4.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (-9.2,'kcal/mol','+|-',1.3),
S298 = (7.2,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1875,
label = "O2s-CsN3d",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 N3d u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1847,
label = "O2s-Cs(N3dOd)",
group =
"""
1 * O2s u0 {2,S} {4,S}
2 N3d u0 {1,S} {3,D}
3 O2d u0 {2,D}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.6,11.3,11.9,12.6,13.6,14.3,14.8],'cal/(mol*K)','+|-',[0.8,0.8,0.8,0.8,0.8,0.8,0.8]),
H298 = (-4.8,'kcal/mol','+|-',1.1),
S298 = (40,'cal/(mol*K)','+|-',1),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1877,
label = "O2s-CdN3d",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cd u0 {1,S}
3 N3d u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1878,
label = "O2s-(Cd-Cd)(N3dOd)",
group =
"""
1 Cd u0 {2,S} {4,D} {5,S}
2 * O2s u0 {1,S} {3,S}
3 N3d u0 {2,S} {6,D}
4 Cd u0 {1,D}
5 R u0 {1,S}
6 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.1,11.7,12.2,12.7,13.5,14.1,14.9],'cal/(mol*K)','+|-',[0.7,0.7,0.7,0.7,0.7,0.7,0.7]),
H298 = (-5.3,'kcal/mol','+|-',0.9),
S298 = (39.5,'cal/(mol*K)','+|-',0.9),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1876,
label = "O2s-CsN5d",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 N5dc u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1848,
label = "O2s-Cs(N5dOdOs)",
group =
"""
1 N5dc u0 {2,S} {3,D} {4,S}
2 * O2s u0 {1,S} {5,S}
3 O2d u0 {1,D}
4 O2s u0 {1,S}
5 Cs u0 {2,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.2,13.9,15.4,16.6,18.4,19.3,19.9],'cal/(mol*K)','+|-',[0.8,0.8,0.8,0.8,0.8,0.8,0.8]),
H298 = (-19.1,'kcal/mol','+|-',1.1),
S298 = (45.3,'cal/(mol*K)','+|-',1),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1879,
label = "O2s-CdN5d",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cd u0 {1,S}
3 N5dc u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1880,
label = "O2s-(Cd-CdHH)(N5dOdOs)",
group =
"""
1 N5dc u0 {3,S} {4,D} {5,S}
2 Cd u0 {3,S} {6,D} {7,S}
3 * O2s u0 {1,S} {2,S}
4 O2d u0 {1,D}
5 O2s u0 {1,S}
6 Cd u0 {2,D}
7 R u0 {2,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([12.4,14.2,15.7,16.9,18.5,19.3,20.1],'cal/(mol*K)','+|-',[0.8,0.8,0.8,0.8,0.8,0.8,0.8]),
H298 = (-18.4,'kcal/mol','+|-',1.1),
S298 = (45.4,'cal/(mol*K)','+|-',1),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1936,
label = "O2s-ON",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 N u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1881,
label = "O2s-OsN3s",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 N3s u0 {1,S}
3 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.3,4.9,5.6,6.3,7,7.1,6.5],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (5.3,'kcal/mol','+|-',1.3),
S298 = (6.9,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1882,
label = "O2s-OsN3d",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 N3d u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1883,
label = "O2s-O2s(N3dOd)",
group =
"""
1 * O2s u0 {2,S} {4,S}
2 N3d u0 {1,S} {3,D}
3 O2d u0 {2,D}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.7,12.9,13.6,14.2,15,15.5,16],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (15.2,'kcal/mol','+|-',1.3),
S298 = (40.7,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1937,
label = "O2s-NN",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 N u0 {1,S}
3 N u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1884,
label = "O2s-N3sN3s",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 N3s u0 {1,S}
3 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.8,4.6,5.1,5.2,5.2,4.9,4.3],'cal/(mol*K)','+|-',[1.6,1.6,1.6,1.6,1.6,1.6,1.6]),
H298 = (5.7,'kcal/mol','+|-',2.2),
S298 = (6.8,'cal/(mol*K)','+|-',2.1),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1885,
label = "O2s-N3sN3d",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 N3s u0 {1,S}
3 N3d u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1886,
label = "O2s-N3s(N3dOd)",
group =
"""
1 * O2s u0 {2,S} {4,S}
2 N3d u0 {1,S} {3,D}
3 O2d u0 {2,D}
4 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.2,11.5,12.4,13,13.9,14.3,14.8],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (10.8,'kcal/mol','+|-',1.3),
S298 = (40.8,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1098,
label = "O2s-HH",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 H u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.03,8.19,8.42,8.68,9.26,9.86,11.26],'cal/(mol*K)'),
H298 = (-57.8,'kcal/mol','+|-',0.01),
S298 = (46.51,'cal/(mol*K)','+|-',0.002),
),
shortDesc = u"""O-HH WATER. !!!Using NIST value for H2O, S(group) = S(H2O) + Rln(2)""",
longDesc =
u"""
""",
)
entry(
index = 1099,
label = "O2s-OsH",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.21,5.72,6.17,6.66,7.15,7.61,8.43],'cal/(mol*K)','+|-',[0.07,0.07,0.07,0.07,0.07,0.07,0.07]),
H298 = (-16.3,'kcal/mol','+|-',0.14),
S298 = (27.83,'cal/(mol*K)','+|-',0.07),
),
shortDesc = u"""O-OH SANDIA 1/2*H2O2""",
longDesc =
u"""
""",
)
entry(
index = 1100,
label = "O2s-OsOs",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2.2,3.64,4.2,4.34,4.62,4.9,4.9],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (8.85,'kcal/mol','+|-',0.16),
S298 = (9.4,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""O-OO LAY 1997=20 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1101,
label = "O2s-CH",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 C u0 {1,S}
3 H u0 {1,S}
""",
thermo = u'O2s-CsH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1102,
label = "O2s-CtH",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Ct u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.3,4.5,4.82,5.23,6.02,6.61,7.44],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-37.9,'kcal/mol','+|-',0.16),
S298 = (29.1,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""O-CtH BENSON (Assigned O-CsH)""",
longDesc =
u"""
""",
)
entry(
index = 1103,
label = "O2s-CdsH",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 [Cd,CO] u0 {1,S}
3 H u0 {1,S}
""",
thermo = u'O2s-(Cds-Cd)H',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1104,
label = "O2s-(Cds-O2d)H",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 CO u0 {1,S} {4,D}
3 H u0 {1,S}
4 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([19.07,19.8,20.85,22.07,24.57,26.95,31.66],'J/(mol*K)','+|-',[2.54,2.54,2.54,2.54,2.54,2.54,2.54]),
H298 = (-165.2,'kJ/mol','+|-',2.16),
S298 = (125.32,'J/(mol*K)','+|-',2.96),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1105,
label = "O2s-(Cds-Cd)H",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cd u0 {1,S} {4,D}
3 H u0 {1,S}
4 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([24.6,30.3,32.52,33.15,33.29,33.55,34.97],'J/(mol*K)','+|-',[4.18,4.18,4.18,4.18,4.18,4.18,4.18]),
H298 = (-188.1,'kJ/mol','+|-',3.56),
S298 = (106.3,'J/(mol*K)','+|-',4.87),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1106,
label = "O2s-CsH",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([19.07,19.8,20.85,22.07,24.57,26.95,31.66],'J/(mol*K)','+|-',[2.54,2.54,2.54,2.54,2.54,2.54,2.54]),
H298 = (-165.2,'kJ/mol','+|-',2.16),
S298 = (125.32,'J/(mol*K)','+|-',2.96),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1107,
label = "O2s-CbH",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cb u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.3,4.5,4.82,5.23,6.02,6.61,7.44],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-37.9,'kcal/mol','+|-',0.16),
S298 = (29.1,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""O-CbH BENSON (Assigned O-CsH)""",
longDesc =
u"""
""",
)
entry(
index = 1460,
label = "O2s-CSH",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 CS u0 {1,S} {4,D}
3 H u0 {1,S}
4 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.98,8.35,9.48,10.38,11.59,12.26,12.99],'cal/(mol*K)'),
H298 = (-31.38,'kcal/mol'),
S298 = (32.08,'cal/(mol*K)'),
),
shortDesc = u"""CAC calc 1D-HR""",
longDesc =
u"""
""",
)
entry(
index = 1108,
label = "O2s-OsC",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 C u0 {1,S}
""",
thermo = u'O2s-OsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1109,
label = "O2s-OsCt",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.9,4.31,4.6,4.84,5.32,5.8,5.8],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (7,'kcal/mol','+|-',0.3),
S298 = (10.8,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""O-OCb Hf JWB plot S,Cp assigned O/O/Cd !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1110,
label = "O2s-OsCds",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 [Cd,CO] u0 {1,S}
""",
thermo = u'O2s-O2s(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1111,
label = "O2s-O2s(Cds-O2d)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 CO u0 {1,S} {4,D}
3 O2s u0 {1,S}
4 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.53,5.02,5.79,6.08,6.54,6.49,6.49],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (-23.22,'kcal/mol','+|-',0.3),
S298 = (9.11,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""O-OCO jwl cbsQ 99 cqcho=20 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1112,
label = "O2s-O2s(Cds-Cd)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cd u0 {1,S} {4,D}
3 O2s u0 {1,S}
4 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.5,3.87,3.95,4.15,4.73,4.89,4.89],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (1.64,'kcal/mol','+|-',0.3),
S298 = (10.12,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""O-OCd WESTMORELAND S,Cp LAY'9405 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1113,
label = "O2s-OsCs",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.9,4.31,4.6,4.84,5.32,5.8,5.8],'cal/(mol*K)','+|-',[0.15,0.15,0.15,0.15,0.15,0.15,0.15]),
H298 = (-5.4,'kcal/mol','+|-',0.3),
S298 = (8.54,'cal/(mol*K)','+|-',0.15),
),
shortDesc = u"""O-OCs LAY 1997 !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1114,
label = "O2s-OsCb",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 O2s u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = u'O2s-O2s(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1115,
label = "O2s-CC",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 C u0 {1,S}
3 C u0 {1,S}
""",
thermo = u'O2s-(Cds-Cd)(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1116,
label = "O2s-CtCt",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
""",
thermo = u'O2s-(Cds-Cd)(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1117,
label = "O2s-CtCds",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Ct u0 {1,S}
3 [Cd,CO] u0 {1,S}
""",
thermo = u'O2s-(Cds-Cd)(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1118,
label = "O2s-Ct(Cds-O2d)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 CO u0 {1,S} {4,D}
3 Ct u0 {1,S}
4 O2d u0 {2,D}
""",
thermo = u'O2s-(Cds-Cd)(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1119,
label = "O2s-Ct(Cds-Cd)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cd u0 {1,S} {4,D}
3 Ct u0 {1,S}
4 C u0 {2,D}
""",
thermo = u'O2s-(Cds-Cd)(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1120,
label = "O2s-CtCs",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Ct u0 {1,S}
3 Cs u0 {1,S}
""",
thermo = u'O2s-Cs(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1873,
label = "O2s-Cs(CtN3t)",
group =
"""
1 * O2s u0 {2,S} {4,S}
2 Ct u0 {1,S} {3,T}
3 N3t u0 {2,T}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([9.1,9.8,10.6,11.2,12.3,13,13.8],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (10,'kcal/mol','+|-',1.3),
S298 = (39.1,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1121,
label = "O2s-CtCb",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Ct u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = u'O2s-(Cds-Cd)(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1122,
label = "O2s-CdsCds",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 [Cd,CO] u0 {1,S}
3 [Cd,CO] u0 {1,S}
""",
thermo = u'O2s-(Cds-Cd)(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1123,
label = "O2s-(Cds-O2d)(Cds-O2d)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 CO u0 {1,S} {4,D}
3 CO u0 {1,S} {5,D}
4 O2d u0 {2,D}
5 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([18.4,11.55,6.97,3.72,-0.53,-2.57,-1.41],'J/(mol*K)','+|-',[6.51,6.51,6.51,6.51,6.51,6.51,6.51]),
H298 = (-46.4,'kJ/mol','+|-',5.54),
S298 = (80.8,'J/(mol*K)','+|-',7.59),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1124,
label = "O2s-(Cds-O2d)(Cds-Cd)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 CO u0 {1,S} {5,D}
3 Cd u0 {1,S} {4,D}
4 C u0 {3,D}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([20.02,19.61,18.5,17.71,17.02,16.49,15.33],'J/(mol*K)','+|-',[8.17,8.17,8.17,8.17,8.17,8.17,8.17]),
H298 = (-100.6,'kJ/mol','+|-',6.96),
S298 = (38.43,'J/(mol*K)','+|-',9.53),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1125,
label = "O2s-(Cds-Cd)(Cds-Cd)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cd u0 {1,S}
3 Cd u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.4,3.7,3.7,3.8,4.4,4.6,4.8],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-19.61,'kcal/mol','+|-',0.19),
S298 = (10,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""O-CdCd BOZZELLI""",
longDesc =
u"""
""",
)
entry(
index = 1126,
label = "O2s-CdsCs",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 [Cd,CO] u0 {1,S}
3 Cs u0 {1,S}
""",
thermo = u'O2s-Cs(Cds-Cd)',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1127,
label = "O2s-Cs(Cds-O2d)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 CO u0 {1,S} {4,D}
3 Cs u0 {1,S}
4 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.49,9.94,9.96,10.7,12.71,14.71,18],'J/(mol*K)','+|-',[3.15,3.15,3.15,3.15,3.15,3.15,3.15]),
H298 = (-102.2,'kJ/mol','+|-',2.69),
S298 = (45.71,'J/(mol*K)','+|-',3.68),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1128,
label = "O2s-Cs(Cds-Cd)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cd u0 {1,S} {4,D}
3 Cs u0 {1,S}
4 C u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([19.07,23.32,25.26,25.92,25.5,24.52,22.72],'J/(mol*K)','+|-',[3.47,3.47,3.47,3.47,3.47,3.47,3.47]),
H298 = (-123.9,'kJ/mol','+|-',2.96),
S298 = (18.91,'J/(mol*K)','+|-',4.05),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = -1,
label = "O2s-CdsCb",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 [Cd,CO] u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "O2s-Cb(Cds-O2d)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 CO u0 {1,S} {4,D}
3 Cb u0 {1,S}
4 O2d u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "O2s-Cb(Cds-Cd)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cd u0 {1,S} {4,D}
3 Cb u0 {1,S}
4 C u0 {2,D}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1129,
label = "O2s-CsCs",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([14.7,13.4,13.58,14.54,16.71,18.29,20.17],'J/(mol*K)','+|-',[2.44,2.44,2.44,2.44,2.44,2.44,2.44]),
H298 = (-98.6,'kJ/mol','+|-',2.08),
S298 = (38.61,'J/(mol*K)','+|-',2.85),
),
shortDesc = u"""\Derived from CBS-QB3 calculation with 1DHR treatment""",
longDesc =
u"""
Derived using calculations at B3LYP/6-311G(d,p)/CBS-QB3 level of theory. 1DH-rotors
optimized at the B3LYP/6-31G(d).Paraskevas et al, Chem. Eur. J. 2013, 19, 16431-16452,
DOI: 10.1002/chem.201301381
""",
)
entry(
index = 1130,
label = "O2s-CsCb",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.4,3.7,3.7,3.8,4.4,4.6,4.6],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-22.6,'kcal/mol','+|-',0.19),
S298 = (9.7,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""O-CbCs REID, PRAUSNITZ and SHERWOOD !!!WARNING! Cp1500 value taken as Cp1000""",
longDesc =
u"""
""",
)
entry(
index = 1131,
label = "O2s-CbCb",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([1.19,-0.24,-0.72,-0.51,0.43,1.36,1.75],'cal/(mol*K)','+|-',[0.1,0.1,0.1,0.1,0.1,0.1,0.1]),
H298 = (-18.77,'kcal/mol','+|-',0.19),
S298 = (13.59,'cal/(mol*K)','+|-',0.1),
),
shortDesc = u"""O-CbCb CHERN 1/97 Hf PEDLEY, Mopac""",
longDesc =
u"""
""",
)
entry(
index = 1461,
label = "O2s-Cs(Cds-S2d)",
group =
"""
1 * O2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 CS u0 {1,S} {4,D}
4 S2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.56,6.31,7,7.61,8.52,8.99,9.29],'cal/(mol*K)'),
H298 = (-14.54,'kcal/mol'),
S298 = (10.02,'cal/(mol*K)'),
),
shortDesc = u"""CAC CBS-QB3 1D-HR""",
longDesc =
u"""
""",
)
entry(
index = 1600,
label = "Si",
group =
"""
1 * Si u0
""",
thermo = u'Cs-HHHH',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1400,
label = "S",
group =
"""
1 * S u0
""",
thermo = u'S2s-CsCs',
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "S2d",
group =
"""
1 * S2d u0
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1160,
label = "S2d-Cd",
group =
"""
1 * S2d u0 {2,D}
2 CS u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1161,
label = "S2d-S2d",
group =
"""
1 * S2d u0 {2,D}
2 S2d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.9,4.08,4.2,4.27,4.35,4.39,4.43],'cal/(mol*K)'),
H298 = (22.82,'kcal/mol'),
S298 = (26.89,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2009""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "S2s",
group =
"""
1 * S2s u0
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1132,
label = "S2s-HH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 H u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.15,8.48,8.85,9.26,10.08,10.82,12.1],'cal/(mol*K)'),
H298 = (-5.37,'kcal/mol'),
S298 = (50.52,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "S2s-CH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 C u0 {1,S}
3 H u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1133,
label = "S2s-CsH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.17,6.22,6.4,6.65,7.18,7.65,8.45],'cal/(mol*K)'),
H298 = (5.05,'kcal/mol'),
S298 = (33.68,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1134,
label = "S2s-CdH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cd u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.01,6.82,7.28,7.55,7.9,8.18,8.7],'cal/(mol*K)'),
H298 = (4.19,'kcal/mol'),
S298 = (32.23,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1135,
label = "S2s-CtH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Ct u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.22,6.87,7.26,7.55,7.94,8.24,8.73],'cal/(mol*K)'),
H298 = (6.27,'kcal/mol'),
S298 = (31.59,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1136,
label = "S2s-CbH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cb u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.65,7.07,7.33,7.5,7.79,8.05,8.53],'cal/(mol*K)'),
H298 = (3.91,'kcal/mol'),
S298 = (31.98,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1462,
label = "S2s-COH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 CO u0 {1,S} {4,D}
3 H u0 {1,S}
4 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.05,9.1,9.95,10.65,11.62,12.26,13.25],'cal/(mol*K)'),
H298 = (-21.06,'kcal/mol'),
S298 = (35.41,'cal/(mol*K)'),
),
shortDesc = u"""CAC calc 1D-HR""",
longDesc =
u"""
""",
)
entry(
index = 1153,
label = "S2s-C=SH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 CS u0 {1,S} {4,D}
3 H u0 {1,S}
4 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.01,6.82,7.28,7.55,7.9,8.18,8.7],'cal/(mol*K)'),
H298 = (4.19,'kcal/mol'),
S298 = (32.23,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1147,
label = "S2s-SsH",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 S2s u0 {1,S}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.75,6.48,7.02,7.43,8.03,8.43,9],'cal/(mol*K)'),
H298 = (1.97,'kcal/mol'),
S298 = (31.73,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1152,
label = "S2s-SsSs",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 S2s u0 {1,S}
3 S2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.19,6.32,6.38,6.44,6.47,6.39,5.95],'cal/(mol*K)'),
H298 = (3.03,'kcal/mol'),
S298 = (11.18,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "S2s-SsC",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 S2s u0 {1,S}
3 C u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1148,
label = "S2s-SsCs",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 S2s u0 {1,S}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.59,5.73,5.79,5.8,5.74,5.65,5.43],'cal/(mol*K)'),
H298 = (6.99,'kcal/mol'),
S298 = (12.61,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1149,
label = "S2s-SsCd",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 S2s u0 {1,S}
3 Cd u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.41,5.92,6.11,6.14,5.98,5.74,5.25],'cal/(mol*K)'),
H298 = (7.62,'kcal/mol'),
S298 = (12.13,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1150,
label = "S2s-SsCt",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 S2s u0 {1,S}
3 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.3,5.8,5.94,5.94,5.77,5.57,5.24],'cal/(mol*K)'),
H298 = (11.93,'kcal/mol'),
S298 = (12.73,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1151,
label = "S2s-SsCb",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 S2s u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.71,5.93,5.98,5.92,5.67,5.38,4.78],'cal/(mol*K)'),
H298 = (7.09,'kcal/mol'),
S298 = (11.38,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1159,
label = "S2s-C=SSs",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 S2s u0 {1,S}
3 CS u0 {1,S} {4,D}
4 S2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.25,5.49,5.65,5.74,5.74,5.65,5.37],'cal/(mol*K)'),
H298 = (7.9,'kcal/mol'),
S298 = (13.34,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = -1,
label = "S2s-CC",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 C u0 {1,S}
3 C u0 {1,S}
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1137,
label = "S2s-CsCs",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.8,5.76,5.63,5.51,5.3,5.18,5.07],'cal/(mol*K)'),
H298 = (11.41,'kcal/mol'),
S298 = (13.72,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1138,
label = "S2s-CsCd",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 Cd u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.68,6.98,6.89,6.65,6.16,5.79,5.33],'cal/(mol*K)'),
H298 = (9.83,'kcal/mol'),
S298 = (11.01,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1463,
label = "S2s-CsCO",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 CO u0 {1,S} {4,D}
4 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.78,6.73,7.6,8.34,9.31,9.77,10.14],'cal/(mol*K)'),
H298 = (-15.33,'kcal/mol'),
S298 = (11.11,'cal/(mol*K)'),
),
shortDesc = u"""CAC CBS-QB3 1dhr calc""",
longDesc =
u"""
""",
)
entry(
index = 1139,
label = "S2s-CsCt",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.45,5.71,5.77,5.73,5.57,5.42,5.2],'cal/(mol*K)'),
H298 = (12.03,'kcal/mol'),
S298 = (13.23,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1140,
label = "S2s-CsCb",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.62,5.74,5.7,5.6,5.38,5.22,5],'cal/(mol*K)'),
H298 = (10.51,'kcal/mol'),
S298 = (12.6,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1141,
label = "S2s-CdCd",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cd u0 {1,S}
3 Cd u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.82,6.48,6.62,6.51,6.09,5.74,5.25],'cal/(mol*K)'),
H298 = (10.56,'kcal/mol'),
S298 = (12.24,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1142,
label = "S2s-CdCt",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cd u0 {1,S}
3 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.34,5.99,6.17,6.13,5.88,5.63,5.27],'cal/(mol*K)'),
H298 = (12.84,'kcal/mol'),
S298 = (12.07,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1143,
label = "S2s-CdCb",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cd u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.91,6.42,6.51,6.38,6,5.65,5.07],'cal/(mol*K)'),
H298 = (10.23,'kcal/mol'),
S298 = (11.93,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1144,
label = "S2s-CtCt",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Ct u0 {1,S}
3 Ct u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.61,5.28,5.46,5.47,5.37,5.27,5.11],'cal/(mol*K)'),
H298 = (19.93,'kcal/mol'),
S298 = (13.38,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1145,
label = "S2s-CtCb",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Ct u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.6,5.94,5.94,5.82,5.56,5.35,5.06],'cal/(mol*K)'),
H298 = (13.27,'kcal/mol'),
S298 = (11.87,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1146,
label = "S2s-CbCb",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.27,5.7,5.8,5.74,5.53,5.35,5.09],'cal/(mol*K)'),
H298 = (10.52,'kcal/mol'),
S298 = (12.32,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1154,
label = "S2s-C=SCs",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cs u0 {1,S}
3 CS u0 {1,S} {4,D}
4 S2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.43,5.15,5.65,5.92,6.04,5.97,5.72],'cal/(mol*K)'),
H298 = (6.87,'kcal/mol'),
S298 = (11.81,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1156,
label = "S2s-C=SCt",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 CS u0 {1,S} {4,D}
3 Ct u0 {1,S}
4 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.39,4.92,5.17,5.28,5.33,5.29,5.19],'cal/(mol*K)'),
H298 = (15.16,'kcal/mol'),
S298 = (14.06,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1158,
label = "S2s-C=SC=S",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 CS u0 {1,S} {4,D}
3 CS u0 {1,S} {5,D}
4 S2d u0 {2,D}
5 S2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.98,5.28,5.67,6.04,6.51,6.52,5.77],'cal/(mol*K)'),
H298 = (12.91,'kcal/mol'),
S298 = (12.96,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1155,
label = "S2s-C=SCd",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 Cd u0 {1,S}
3 CS u0 {1,S} {4,D}
4 S2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.13,6.41,7.01,7.14,6.87,6.48,5.84],'cal/(mol*K)'),
H298 = (7.78,'kcal/mol'),
S298 = (10.23,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1157,
label = "S2s-C=SCb",
group =
"""
1 * S2s u0 {2,S} {3,S}
2 CS u0 {1,S} {4,D}
3 Cb u0 {1,S}
4 S2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.08,5.5,5.68,5.7,5.59,5.42,5.05],'cal/(mol*K)'),
H298 = (10.76,'kcal/mol'),
S298 = (13.05,'cal/(mol*K)'),
),
shortDesc = u"""CBS-QB3 GA 1D-HR Aaron Vandeputte 2010""",
longDesc =
u"""
""",
)
entry(
index = 1887,
label = "N",
group =
"""
1 * [N1dc,N3s,N3d,N3t,N5sc,N5dc,N5ddc,N5tc] u0
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1922,
label = "N1dc",
group =
"""
1 * N1dc u0 p2
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1888,
label = "N3s",
group =
"""
1 * N3s u0
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1938,
label = "N3s-CHH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 C u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1808,
label = "N3s-CsHH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.72,6.51,7.32,8.07,9.41,10.47,12.28],'cal/(mol*K)'),
H298 = (4.8,'kcal/mol'),
S298 = (29.71,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1817,
label = "N3s-CbHH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cb u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.72,6.51,7.32,8.07,9.41,10.47,12.28],'cal/(mol*K)'),
H298 = (4.8,'kcal/mol'),
S298 = (29.71,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1825,
label = "N3s-(CO)HH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 CO u0 {1,S} {5,D}
3 H u0 {1,S}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.07,5.74,7.13,8.29,9.96,11.22,14.37],'cal/(mol*K)'),
H298 = (-14.9,'kcal/mol'),
S298 = (-24.69,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1889,
label = "N3s-CdHH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cd u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.7,6.5,7.3,8.1,9.4,10.5,12.3],'cal/(mol*K)'),
H298 = (4.8,'kcal/mol'),
S298 = (29.7,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1938,
label = "N3s-CCH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1809,
label = "N3s-CsCsH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.2,5.21,6.13,6.83,7.9,8.65,9.55],'cal/(mol*K)'),
H298 = (15.4,'kcal/mol'),
S298 = (8.94,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1818,
label = "N3s-CbCsH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (14.9,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1820,
label = "N3s-CbCbH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cb u0 {1,S}
3 Cb u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (16.3,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1826,
label = "N3s-(CO)CsH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 CO u0 {1,S} {5,D}
3 Cs u0 {1,S}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (-4.4,'kcal/mol'),
S298 = (3.9,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1828,
label = "N3s-(CO)CbH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 CO u0 {1,S} {5,D}
3 Cb u0 {1,S}
4 H u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0.4,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1829,
label = "N3s-(CO)(CO)H",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 CO u0 {1,S} {5,D}
3 CO u0 {1,S} {6,D}
4 H u0 {1,S}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (-18.5,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1894,
label = "N3s-(CtN3t)CsH",
group =
"""
1 * N3s u0 {2,S} {4,S} {5,S}
2 Ct u0 {1,S} {3,T}
3 N3t u0 {2,T}
4 Cs u0 {1,S}
5 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.3,11.6,12.8,13.9,15.5,16.7,18.3],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (44.1,'kcal/mol','+|-',1.3),
S298 = (40.7,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1899,
label = "N3s-(CdCd)CsH",
group =
"""
1 * N3s u0 {2,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D} {4,S}
3 Cd u0 {2,D}
4 R u0 {2,S}
5 Cs u0 {1,S}
6 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.8,6.1,6.4,6.7,7.5,8.1,9.1],'cal/(mol*K)','+|-',[1.3,1.3,1.3,1.3,1.3,1.3,1.3]),
H298 = (15.3,'kcal/mol','+|-',1.9),
S298 = (8.7,'cal/(mol*K)','+|-',1.7),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1939,
label = "N3s-CCC",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 C u0 {1,S}
3 C u0 {1,S}
4 C u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1810,
label = "N3s-CsCsCs",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.48,4.56,5.43,5.97,6.56,6.67,6.5],'cal/(mol*K)'),
H298 = (24.4,'kcal/mol'),
S298 = (-13.46,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1819,
label = "N3s-CbCsCs",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cb u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (26.2,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1827,
label = "N3s-(CO)CsCs",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 CO u0 {1,S} {5,D}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1830,
label = "N3s-(CO)(CO)Cs",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 CO u0 {1,S} {5,D}
3 CO u0 {1,S} {6,D}
4 Cs u0 {1,S}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (-5.9,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1831,
label = "N3s-(CO)(CO)Cb",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 CO u0 {1,S} {5,D}
3 CO u0 {1,S} {6,D}
4 Cb u0 {1,S}
5 O2d u0 {2,D}
6 O2d u0 {3,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (-0.5,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1895,
label = "N3s-(CtN3t)CsCs",
group =
"""
1 * N3s u0 {2,S} {4,S} {5,S}
2 Ct u0 {1,S} {3,T}
3 N3t u0 {2,T}
4 Cs u0 {1,S}
5 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([8.6,9.6,10.5,11.4,12.9,13.8,14.8],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (53.3,'kcal/mol','+|-',1.3),
S298 = (21,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1898,
label = "N3s-(CdCd)CsCs",
group =
"""
1 * N3s u0 {2,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D} {4,S}
3 Cd u0 {2,D}
4 R u0 {2,S}
5 Cs u0 {1,S}
6 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2.8,2.9,3.3,3.7,4.6,5,5.5],'cal/(mol*K)','+|-',[1.3,1.3,1.3,1.3,1.3,1.3,1.3]),
H298 = (25.9,'kcal/mol','+|-',1.9),
S298 = (-11,'cal/(mol*K)','+|-',1.7),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1811,
label = "N3s-N3sHH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 H u0 {1,S}
3 H u0 {1,S}
4 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.1,7.38,8.43,9.27,10.54,11.52,13.19],'cal/(mol*K)'),
H298 = (11.4,'kcal/mol'),
S298 = (29.13,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1940,
label = "N3s-NCH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 N u0 {1,S}
3 C u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1812,
label = "N3s-N3sCsH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 H u0 {1,S}
4 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.82,5.8,6.5,7,7.8,8.3,9],'cal/(mol*K)'),
H298 = (20.9,'kcal/mol'),
S298 = (9.61,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1814,
label = "N3s-N3sCbH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 N3s u0 {1,S}
3 Cb u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (22.1,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1897,
label = "N3s-CsH(N3dOd)",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 H u0 {1,S}
4 N3d u0 {1,S} {5,D}
5 O2d u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([10.4,11.9,13.4,14.7,16.6,17.9,19.2],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (25.2,'kcal/mol','+|-',1.3),
S298 = (41.7,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1902,
label = "N3s-CsH(N5dOdOs)",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 H u0 {1,S}
4 N5dc u0 {1,S} {5,D} {6,S}
5 O2d u0 {4,D}
6 O2s u0 {4,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([13.1,15.5,17.6,19.2,21.4,22.8,24.4],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (8.4,'kcal/mol','+|-',1.3),
S298 = (45.3,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1901,
label = "N3s-(CdCd)HN3s",
group =
"""
1 * N3s u0 {2,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D} {4,S}
3 Cd u0 {2,D}
4 R u0 {2,S}
5 H u0 {1,S}
6 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.5,5.4,6.5,7.3,8.5,9.1,9.9],'cal/(mol*K)','+|-',[1.1,1.1,1.1,1.1,1.1,1.1,1.1]),
H298 = (20.5,'kcal/mol','+|-',1.5),
S298 = (6.6,'cal/(mol*K)','+|-',1.4),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1940,
label = "N3s-NCC",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 N u0 {1,S}
3 C u0 {1,S}
4 C u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1893,
label = "N3s-NCsCs",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 N u0 {1,S}
3 Cs u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (29.2,'kcal/mol'),
S298 = (-13.8,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1813,
label = "N3s-CsCsN3s",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.7,4.9,5.8,6.3,6.8,6.8,6.7],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (26.8,'kcal/mol','+|-',1.3),
S298 = (-14.5,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1896,
label = "N3s-CsCs(N3dOd)",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 N3d u0 {1,S} {5,D}
5 O2d u0 {4,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([9.4,10.5,11.5,12.4,13.8,14.6,15.3],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (32.6,'kcal/mol','+|-',1.3),
S298 = (19.3,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1903,
label = "N3s-CsCs(N5dOdOs)",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 N5dc u0 {1,S} {5,D} {6,S}
5 O2d u0 {4,D}
6 O2s u0 {4,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([11.5,13.4,15.2,16.7,18.8,20,21.1],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (16.7,'kcal/mol','+|-',1.3),
S298 = (25.8,'cal/(mol*K)','+|-',1.2),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1941,
label = "N3s-NCdCs",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 N u0 {1,S}
3 Cd u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1900,
label = "N3s-(CdCd)CsN3s",
group =
"""
1 * N3s u0 {2,S} {5,S} {6,S}
2 Cd u0 {1,S} {3,D} {4,S}
3 Cd u0 {2,D}
4 R u0 {2,S}
5 Cs u0 {1,S}
6 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.2,4.2,4.4,4.8,5.4,5.7,6],'cal/(mol*K)','+|-',[1.1,1.1,1.1,1.1,1.1,1.1,1.1]),
H298 = (30.3,'kcal/mol','+|-',1.5),
S298 = (-13.2,'cal/(mol*K)','+|-',1.4),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1891,
label = "N3s-CsHOs",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 H u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([5.2,6.2,7,7.7,8.7,9.4,10.5],'cal/(mol*K)','+|-',[1,1,1,1,1,1,1]),
H298 = (20.4,'kcal/mol','+|-',1.4),
S298 = (8.1,'cal/(mol*K)','+|-',1.3),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1892,
label = "N3s-CsCsOs",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 Cs u0 {1,S}
3 Cs u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.3,5.1,5.7,6.2,7,7.3,7.5],'cal/(mol*K)','+|-',[0.8,0.8,0.8,0.8,0.8,0.8,0.8]),
H298 = (26.6,'kcal/mol','+|-',1.2),
S298 = (-12.7,'cal/(mol*K)','+|-',1.1),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1890,
label = "N3s-OsHH",
group =
"""
1 * N3s u0 {2,S} {3,S} {4,S}
2 O2s u0 {1,S}
3 H u0 {1,S}
4 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([6.1,7.4,8.4,9.3,10.5,11.5,13.2],'cal/(mol*K)'),
H298 = (11.4,'kcal/mol'),
S298 = (29.1,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1904,
label = "N3d",
group =
"""
1 * N3d u0
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1906,
label = "N3d-CdH",
group =
"""
1 * N3d u0 {2,D} {3,S}
2 [Cd,Cdd] u0 {1,D}
3 H u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3,3.5,3.9,4.3,5,5.5,6.4],'cal/(mol*K)'),
H298 = (16.3,'kcal/mol'),
S298 = (13.3,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1815,
label = "N3d-N3dH",
group =
"""
1 * N3d u0 {2,S} {3,D}
2 H u0 {1,S}
3 N3d u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([4.38,4.89,5.44,5.94,6.77,7.42,8.44],'cal/(mol*K)'),
H298 = (25.1,'kcal/mol'),
S298 = (26.8,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1822,
label = "N3d-N3dN3s",
group =
"""
1 * N3d u0 {2,D} {3,S}
2 N3d u0 {1,D}
3 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (23,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1909,
label = "N3d-OdOs",
group =
"""
1 * N3d u0 {2,D} {3,S}
2 O2d u0 {1,D}
3 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1910,
label = "N3d-OdN3s",
group =
"""
1 * N3d u0 {2,D} {3,S}
2 O2d u0 {1,D}
3 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1911,
label = "N3d-CsR",
group =
"""
1 * N3d u0 {2,S} {3,D}
2 Cs u0 {1,S}
3 R!H u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (21.3,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1908,
label = "N3d-OdC",
group =
"""
1 * N3d u0 {2,D} {3,S}
2 O2d u0 {1,D}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1905,
label = "N3d-CdCs",
group =
"""
1 * N3d u0 {2,D} {3,S}
2 [Cd,Cdd] u0 {1,D}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([2,2.2,2.2,2.3,2.5,2.7,2.9],'cal/(mol*K)'),
H298 = (21.3,'kcal/mol'),
S298 = (-6.3,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1907,
label = "N3d-N3dCs",
group =
"""
1 * N3d u0 {2,D} {3,S}
2 N3d u0 {1,D}
3 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([3.4,3.6,3.7,3.9,4.3,4.6,4.9],'cal/(mol*K)'),
H298 = (27,'kcal/mol'),
S298 = (7.2,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1912,
label = "N3d-CbR",
group =
"""
1 * N3d u0 {2,S} {3,D}
2 Cb u0 {1,S}
3 R!H u0 {1,D}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (16.7,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1913,
label = "N5dc",
group =
"""
1 * N5dc u0
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1914,
label = "N5dc-OdOsCs",
group =
"""
1 * N5dc u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 O2s u0 {1,S}
4 Cs u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1915,
label = "N5dc-OdOsCd",
group =
"""
1 * N5dc u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 O2s u0 {1,S}
4 Cd u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1916,
label = "N5dc-OdOsOs",
group =
"""
1 * N5dc u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 O2s u0 {1,S}
4 O2s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1917,
label = "N5dc-OdOsN3s",
group =
"""
1 * N5dc u0 {2,D} {3,S} {4,S}
2 O2d u0 {1,D}
3 O2s u0 {1,S}
4 N3s u0 {1,S}
""",
thermo = ThermoData(
Tdata = ([300,400,500,600,800,1000,1500],'K'),
Cpdata = ([0,0,0,0,0,0,0],'cal/(mol*K)'),
H298 = (0,'kcal/mol'),
S298 = (0,'cal/(mol*K)'),
),
shortDesc = u"""""",
longDesc =
u"""
""",
)
entry(
index = 1918,
label = "N5ddc",
group =
"""
1 * N5ddc u0
""",
thermo = None,
shortDesc = u"""""",
longDesc =
u"""
""",
)
tree(
"""
L1: R
L2: C
L3: Cbf
L4: Cbf-CbCbCbf
L4: Cbf-CbCbfCbf
L4: Cbf-CbfCbfCbf
L3: Cb
L4: Cb-H
L4: Cb-O2s
L4: Cb-S2s
L4: Cb-N3s
L4: Cb-C
L5: Cb-Cs
L5: Cb-Cds
L6: Cb-(Cds-O2d)
L6: Cb-(Cds-Cd)
L7: Cb-(Cds-Cds)
L7: Cb-(Cds-Cdd)
L8: Cb-(Cds-Cdd-O2d)
L8: Cb-(Cds-Cdd-S2d)
L8: Cb-(Cds-Cdd-Cd)
L5: Cb-Ct
L6: Cb-(CtN3t)
L5: Cb-Cb
L5: Cb-C=S
L3: Ct
L4: Ct-CtN3s
L4: Ct-N3tN3s
L4: Ct-CtH
L4: Ct-CtOs
L4: Ct-N3tOs
L4: Ct-CtSs
L4: Ct-N3tC
L5: Ct-N3tCs
L5: Ct-N3tCd
L4: Ct-CtC
L5: Ct-CtCs
L5: Ct-CtCds
L6: Ct-Ct(Cds-O2d)
L6: Ct-Ct(Cds-Cd)
L7: Ct-Ct(Cds-Cds)
L7: Ct-Ct(Cds-Cdd)
L8: Ct-Ct(Cds-Cdd-O2d)
L8: Ct-Ct(Cds-Cdd-S2d)
L8: Ct-Ct(Cds-Cdd-Cd)
L5: Ct-CtCt
L6: Ct-Ct(CtN3t)
L5: Ct-CtCb
L5: Ct-CtC=S
L3: Cdd
L4: Cdd-N3dCd
L4: Cdd-OdOd
L4: Cdd-OdSd
L4: Cdd-SdSd
L4: Cdd-CdOd
L5: Cdd-CdsOd
L5: Cdd-CddOd
L6: Cdd-(Cdd-O2d)O2d
L6: Cdd-(Cdd-Cd)O2d
L4: Cdd-CdSd
L5: Cdd-CdsSd
L5: Cdd-CddSd
L6: Cdd-(Cdd-S2d)S2d
L6: Cdd-(Cdd-Cd)S2d
L4: Cdd-CdCd
L5: Cdd-CddCdd
L6: Cdd-(Cdd-O2d)(Cdd-O2d)
L6: Cdd-(Cdd-S2d)(Cdd-S2d)
L6: Cdd-(Cdd-O2d)(Cdd-Cd)
L6: Cdd-(Cdd-S2d)(Cdd-Cd)
L6: Cdd-(Cdd-Cd)(Cdd-Cd)
L5: Cdd-CddCds
L6: Cdd-(Cdd-O2d)Cds
L6: Cdd-(Cdd-S2d)Cds
L6: Cdd-(Cdd-Cd)Cds
L5: Cdd-CdsCds
L3: Cds
L4: Cds-OdN3sH
L4: Cds-OdN3sCs
L4: Cd-N3dCsCs
L4: Cd-N3dCsH
L4: Cd-N3dHH
L4: Cds-OdHH
L4: Cds-OdOsH
L4: CO-SsH
L4: Cds-OdOsOs
L4: CO-CsSs
L4: CO-OsSs
L4: Cds-OdCH
L5: Cds-OdCsH
L5: Cds-OdCdsH
L6: Cds-O2d(Cds-O2d)H
L6: Cds-O2d(Cds-Cd)H
L7: Cds-O2d(Cds-Cds)H
L7: Cds-O2d(Cds-Cdd)H
L8: Cds-O2d(Cds-Cdd-O2d)H
L8: Cds-O2d(Cds-Cdd-Cd)H
L5: Cds-OdCtH
L5: Cds-OdCbH
L4: Cds-OdCOs
L5: Cds-OdCsOs
L5: Cds-OdCdsOs
L6: Cds-O2d(Cds-O2d)O2s
L6: Cds-O2d(Cds-Cd)O2s
L7: Cds-O2d(Cds-Cds)O2s
L7: Cds-O2d(Cds-Cdd)O2s
L8: Cds-O2d(Cds-Cdd-O2d)O2s
L8: Cds-O2d(Cds-Cdd-Cd)O2s
L5: Cds-OdCtOs
L5: Cds-OdCbOs
L4: Cds-OdCC
L5: Cds-OdCsCs
L5: Cds-OdCdsCs
L6: Cds-O2d(Cds-O2d)Cs
L6: Cds-O2d(Cds-Cd)Cs
L7: Cds-O2d(Cds-Cds)Cs
L7: Cds-O2d(Cds-Cdd)Cs
L8: Cds-O2d(Cds-Cdd-O2d)Cs
L8: Cds-O2d(Cds-Cdd-Cd)Cs
L5: Cds-OdCdsCds
L6: Cds-O2d(Cds-O2d)(Cds-O2d)
L6: Cds-O2d(Cds-Cd)(Cds-O2d)
L7: Cds-O2d(Cds-Cds)(Cds-O2d)
L7: Cds-O2d(Cds-Cdd)(Cds-O2d)
L8: Cds-O2d(Cds-Cdd-O2d)(Cds-O2d)
L8: Cds-O2d(Cds-Cdd-Cd)(Cds-O2d)
L6: Cds-O2d(Cds-Cd)(Cds-Cd)
L7: Cds-O2d(Cds-Cds)(Cds-Cds)
L7: Cds-O2d(Cds-Cdd)(Cds-Cds)
L8: Cds-O2d(Cds-Cdd-O2d)(Cds-Cds)
L8: Cds-O2d(Cds-Cdd-Cd)(Cds-Cds)
L7: Cds-O2d(Cds-Cdd)(Cds-Cdd)
L8: Cds-O2d(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-O2d)
L8: Cds-O2d(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L5: Cds-OdCtCs
L5: Cds-OdCtCds
L6: Cds-OdCt(Cds-O2d)
L6: Cds-OdCt(Cds-Cd)
L7: Cds-OdCt(Cds-Cds)
L7: Cds-OdCt(Cds-Cdd)
L8: Cds-OdCt(Cds-Cdd-O2d)
L8: Cds-OdCt(Cds-Cdd-Cd)
L5: Cds-OdCtCt
L5: Cds-OdCbCs
L5: Cds-OdCbCds
L6: Cds-OdCb(Cds-O2d)
L6: Cds-OdCb(Cds-Cd)
L7: Cds-OdCb(Cds-Cds)
L7: Cds-OdCb(Cds-Cdd)
L8: Cds-OdCb(Cds-Cdd-O2d)
L8: Cds-OdCb(Cds-Cdd-Cd)
L5: Cds-OdCbCt
L5: Cds-OdCbCb
L4: Cds-CdHH
L5: Cds-CdsHH
L5: Cds-CddHH
L6: Cds-(Cdd-O2d)HH
L6: Cds-(Cdd-S2d)HH
L6: Cds-(Cdd-Cd)HH
L4: Cds-CdOsH
L5: Cds-CdsOsH
L5: Cds-CddOsH
L6: Cds-(Cdd-O2d)OsH
L6: Cds-(Cdd-Cd)OsH
L4: Cds-CdSsH
L5: Cds-CdsSsH
L5: Cds-CddSsH
L6: Cds-(Cdd-S2d)SsH
L6: Cds-(Cdd-Cd)SsH
L4: Cds-CdOsOs
L5: Cds-CdsOsOs
L5: Cds-CddOsOs
L6: Cds-(Cdd-O2d)OsOs
L6: Cds-(Cdd-Cd)OsOs
L4: Cds-CdSsSs
L5: Cds-CdsSsSs
L5: Cds-CddSsSs
L6: Cds-(Cdd-S2d)SsSs
L6: Cds-(Cdd-Cd)SsSs
L4: Cds-CdCH
L5: Cds-CdsCsH
L5: Cds-CdsCdsH
L6: Cd-Cd(CO)H
L6: Cds-Cds(Cds-Cd)H
L7: Cds-Cds(Cds-Cds)H
L7: Cds-Cds(Cds-Cdd)H
L8: Cd-Cd(CCO)H
L8: Cds-Cds(Cds-Cdd-S2d)H
L8: Cds-Cds(Cds-Cdd-Cd)H
L5: Cds-CdsCtH
L6: Cds-CdsH(CtN3t)
L5: Cds-CdsCbH
L5: Cds-CddCsH
L6: Cds-(Cdd-O2d)CsH
L6: Cds-(Cdd-S2d)CsH
L6: Cds-(Cdd-Cd)CsH
L5: Cds-CddCdsH
L6: Cds-(Cdd-O2d)(Cds-O2d)H
L6: Cds-(Cdd-O2d)(Cds-Cd)H
L7: Cds-(Cdd-O2d)(Cds-Cds)H
L7: Cds-(Cdd-O2d)(Cds-Cdd)H
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)H
L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)H
L6: Cds-(Cdd-S2d)(Cds-Cd)H
L7: Cds-(Cdd-S2d)(Cds-Cds)H
L7: Cds-(Cdd-S2d)(Cds-Cdd)H
L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)H
L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)H
L6: Cds-(Cdd-Cd)(Cds-O2d)H
L6: Cds-(Cdd-Cd)(Cds-Cd)H
L7: Cds-(Cdd-Cd)(Cds-Cds)H
L7: Cds-(Cdd-Cd)(Cds-Cdd)H
L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)H
L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)H
L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)H
L5: Cds-CddCtH
L6: Cds-(Cdd-O2d)CtH
L6: Cds-(Cdd-S2d)CtH
L6: Cds-(Cdd-Cd)CtH
L5: Cds-CddCbH
L6: Cds-(Cdd-O2d)CbH
L6: Cds-(Cdd-S2d)CbH
L6: Cds-(Cdd-Cd)CbH
L5: Cds-(Cdd-Cd)C=SH
L5: Cds-(Cdd-S2d)C=SH
L5: Cds-CdsC=SH
L4: Cds-CdCO
L5: Cds-CdsCdsOs
L6: Cds-Cds(Cds-O2d)O2s
L6: Cds-Cds(Cds-Cd)O2s
L7: Cds-Cds(Cds-Cds)O2s
L7: Cds-Cds(Cds-Cdd)O2s
L8: Cds-Cds(Cds-Cdd-O2d)O2s
L8: Cds-Cds(Cds-Cdd-Cd)O2s
L5: Cds-CdsCtOs
L5: Cds-CdsCbOs
L5: Cds-CddCdsOs
L6: Cds-(Cdd-O2d)(Cds-O2d)O2s
L6: Cds-(Cdd-O2d)(Cds-Cd)O2s
L7: Cds-(Cdd-O2d)(Cds-Cds)O2s
L7: Cds-(Cdd-O2d)(Cds-Cdd)O2s
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)O2s
L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)O2s
L6: Cds-(Cdd-Cd)(Cds-Cd)O2s
L7: Cds-(Cdd-Cd)(Cds-Cds)O2s
L7: Cds-(Cdd-Cd)(Cds-Cdd)O2s
L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)O2s
L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)O2s
L5: Cds-CddCtOs
L6: Cds-(Cdd-O2d)CtOs
L6: Cds-(Cdd-Cd)CtOs
L5: Cds-CddCbOs
L6: Cds-(Cdd-O2d)CbOs
L6: Cds-(Cdd-Cd)CbOs
L5: Cd-CdCsOs
L6: Cds-CdsCsOs
L6: Cds-CddCsOs
L7: Cds-(Cdd-O2d)CsOs
L7: Cds-(Cdd-Cd)CsOs
L4: Cds-CdCS
L5: Cds-CdsCsSs
L5: Cds-CdsCdsSs
L6: Cds-Cds(Cds-Cd)S2s
L7: Cds-Cds(Cds-Cds)S2s
L7: Cds-Cds(Cds-Cdd)S2s
L8: Cds-Cds(Cds-Cdd-S2d)S2s
L8: Cds-Cds(Cds-Cdd-Cd)S2s
L5: Cds-CdsCtSs
L5: Cds-CdsCbSs
L5: Cds-CddCsSs
L6: Cds-(Cdd-S2d)CsSs
L6: Cds-(Cdd-Cd)CsSs
L5: Cds-CddCdsSs
L6: Cds-(Cdd-S2d)(Cds-Cd)S2s
L7: Cds-(Cdd-S2d)(Cds-Cds)S2s
L7: Cds-(Cdd-S2d)(Cds-Cdd)S2s
L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)S2s
L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)S2s
L6: Cds-(Cdd-Cd)(Cds-Cd)S2s
L7: Cds-(Cdd-Cd)(Cds-Cds)S2s
L7: Cds-(Cdd-Cd)(Cds-Cdd)S2s
L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)S2s
L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)S2s
L5: Cds-CddCtSs
L6: Cds-(Cdd-S2d)CtSs
L6: Cds-(Cdd-Cd)CtSs
L5: Cds-CddCbSs
L6: Cds-(Cdd-S2d)CbSs
L6: Cds-(Cdd-Cd)CbSs
L5: Cds-(Cdd-S2d)C=SSs
L5: Cds-CdsC=SSs
L4: Cds-CdCC
L5: Cds-CdsCsCs
L5: Cds-CdsCdsCs
L6: Cd-CdCs(CO)
L6: Cds-Cds(Cds-Cd)Cs
L7: Cds-Cds(Cds-Cds)Cs
L7: Cds-Cds(Cds-Cdd)Cs
L8: Cd-CdCs(CCO)
L8: Cds-Cds(Cds-Cdd-S2d)Cs
L8: Cds-Cds(Cds-Cdd-Cd)Cs
L5: Cds-CdsCdsCds
L6: Cds-Cds(Cds-O2d)(Cds-O2d)
L6: Cds-Cds(Cds-O2d)(Cds-Cd)
L7: Cds-Cds(Cds-O2d)(Cds-Cds)
L7: Cds-Cds(Cds-O2d)(Cds-Cdd)
L8: Cds-Cds(Cds-O2d)(Cds-Cdd-O2d)
L8: Cds-Cds(Cds-O2d)(Cds-Cdd-Cd)
L6: Cds-Cds(Cds-Cd)(Cds-Cd)
L7: Cds-Cds(Cds-Cds)(Cds-Cds)
L7: Cds-Cds(Cds-Cds)(Cds-Cdd)
L8: Cds-Cds(Cds-Cds)(Cds-Cdd-O2d)
L8: Cds-Cds(Cds-Cds)(Cds-Cdd-S2d)
L8: Cds-Cds(Cds-Cds)(Cds-Cdd-Cd)
L7: Cds-Cds(Cds-Cdd)(Cds-Cdd)
L8: Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cds-Cds(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cds-Cds(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L8: Cds-Cds(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L8: Cds-Cds(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L5: Cds-CdsCtCs
L6: Cd-CdCs(CtN3t)
L5: Cds-CdsCtCds
L6: Cds-CdsCt(Cds-O2d)
L6: Cds-CdsCt(Cds-Cd)
L7: Cds-Cds(Cds-Cds)Ct
L7: Cds-Cds(Cds-Cdd)Ct
L8: Cds-Cds(Cds-Cdd-O2d)Ct
L8: Cds-Cds(Cds-Cdd-S2d)Ct
L8: Cds-Cds(Cds-Cdd-Cd)Ct
L5: Cds-CdsCtCt
L6: Cds-Cd(CtN3t)(CtN3t)
L5: Cds-CdsCbCs
L5: Cds-CdsCbCds
L6: Cds-CdsCb(Cds-O2d)
L6: Cds-Cds(Cds-Cd)Cb
L7: Cds-Cds(Cds-Cds)Cb
L7: Cds-Cds(Cds-Cdd)Cb
L8: Cds-Cds(Cds-Cdd-O2d)Cb
L8: Cds-Cds(Cds-Cdd-S2d)Cb
L8: Cds-Cds(Cds-Cdd-Cd)Cb
L5: Cds-CdsCbCt
L5: Cds-CdsCbCb
L5: Cds-CddCsCs
L6: Cds-(Cdd-O2d)CsCs
L6: Cds-(Cdd-S2d)CsCs
L6: Cds-(Cdd-Cd)CsCs
L5: Cds-CddCdsCs
L6: Cds-(Cdd-O2d)(Cds-O2d)Cs
L6: Cds-(Cdd-O2d)(Cds-Cd)Cs
L7: Cds-(Cdd-O2d)(Cds-Cds)Cs
L7: Cds-(Cdd-O2d)(Cds-Cdd)Cs
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cs
L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cs
L6: Cds-(Cdd-S2d)(Cds-Cd)Cs
L7: Cds-(Cdd-S2d)(Cds-Cds)Cs
L7: Cds-(Cdd-S2d)(Cds-Cdd)Cs
L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)Cs
L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)Cs
L6: Cds-(Cdd-Cd)(Cds-Cd)Cs
L7: Cds-(Cdd-Cd)(Cds-Cds)Cs
L7: Cds-(Cdd-Cd)(Cds-Cdd)Cs
L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)Cs
L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)Cs
L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cs
L5: Cds-CddCdsCds
L6: Cds-(Cdd-O2d)(Cds-O2d)(Cds-O2d)
L6: Cds-(Cdd-O2d)(Cds-Cd)(Cds-O2d)
L7: Cds-(Cdd-O2d)(Cds-Cds)(Cds-O2d)
L7: Cds-(Cdd-O2d)(Cds-Cdd)(Cds-O2d)
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-O2d)
L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-O2d)
L6: Cds-(Cdd-O2d)(Cds-Cd)(Cds-Cd)
L7: Cds-(Cdd-O2d)(Cds-Cds)(Cds-Cds)
L7: Cds-(Cdd-O2d)(Cds-Cdd)(Cds-Cds)
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cds)
L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cds)
L7: Cds-(Cdd-O2d)(Cds-Cdd)(Cds-Cdd)
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L6: Cds-(Cdd-Cd)(Cds-O2d)(Cds-O2d)
L6: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cd)
L7: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cds)
L7: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd)
L8: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-O2d)
L8: Cds-(Cdd-Cd)(Cds-O2d)(Cds-Cdd-Cd)
L6: Cds-(Cdd-S2d)(Cds-Cd)(Cds-Cd)
L7: Cds-(Cdd-S2d)(Cds-Cds)(Cds-Cds)
L7: Cds-(Cdd-S2d)(Cds-Cdd)(Cds-Cds)
L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cds)
L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cds)
L7: Cds-(Cdd-S2d)(Cds-Cdd)(Cds-Cdd)
L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L6: Cds-(Cdd-Cd)(Cds-Cd)(Cds-Cd)
L7: Cds-(Cdd-Cd)(Cds-Cds)(Cds-Cds)
L7: Cds-(Cdd-Cd)(Cds-Cdd)(Cds-Cds)
L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cds)
L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cds)
L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cds)
L7: Cds-(Cdd-Cd)(Cds-Cdd)(Cds-Cdd)
L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L5: Cds-CddCtCs
L6: Cds-(Cdd-O2d)CtCs
L6: Cds-(Cdd-S2d)CtCs
L6: Cds-(Cdd-Cd)CtCs
L5: Cds-CddCtCds
L6: Cds-(Cdd-O2d)(Cds-O2d)Ct
L6: Cds-(Cdd-O2d)(Cds-Cd)Ct
L7: Cds-(Cdd-O2d)(Cds-Cds)Ct
L7: Cds-(Cdd-O2d)(Cds-Cdd)Ct
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)Ct
L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)Ct
L6: Cds-(Cdd-S2d)(Cds-Cd)Ct
L7: Cds-(Cdd-S2d)(Cds-Cds)Ct
L7: Cds-(Cdd-S2d)(Cds-Cdd)Ct
L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)Ct
L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)Ct
L6: Cds-(Cdd-Cd)(Cds-Cd)Ct
L7: Cds-(Cdd-Cd)(Cds-Cds)Ct
L7: Cds-(Cdd-Cd)(Cds-Cdd)Ct
L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)Ct
L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)Ct
L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)Ct
L5: Cds-CddCtCt
L6: Cds-(Cdd-O2d)CtCt
L6: Cds-(Cdd-S2d)CtCt
L6: Cds-(Cdd-Cd)CtCt
L5: Cds-CddCbCs
L6: Cds-(Cdd-O2d)CbCs
L6: Cds-(Cdd-S2d)CbCs
L6: Cds-(Cdd-Cd)CbCs
L5: Cds-CddCbCds
L6: Cds-(Cdd-O2d)(Cds-O2d)Cb
L6: Cds-(Cdd-O2d)(Cds-Cd)Cb
L7: Cds-(Cdd-O2d)(Cds-Cds)Cb
L7: Cds-(Cdd-O2d)(Cds-Cdd)Cb
L8: Cds-(Cdd-O2d)(Cds-Cdd-O2d)Cb
L8: Cds-(Cdd-O2d)(Cds-Cdd-Cd)Cb
L6: Cds-(Cdd-S2d)(Cds-Cd)Cb
L7: Cds-(Cdd-S2d)(Cds-Cds)Cb
L7: Cds-(Cdd-S2d)(Cds-Cdd)Cb
L8: Cds-(Cdd-S2d)(Cds-Cdd-S2d)Cb
L8: Cds-(Cdd-S2d)(Cds-Cdd-Cd)Cb
L6: Cds-(Cdd-Cd)(Cds-Cd)Cb
L7: Cds-(Cdd-Cd)(Cds-Cds)Cb
L7: Cds-(Cdd-Cd)(Cds-Cdd)Cb
L8: Cds-(Cdd-Cd)(Cds-Cdd-O2d)Cb
L8: Cds-(Cdd-Cd)(Cds-Cdd-S2d)Cb
L8: Cds-(Cdd-Cd)(Cds-Cdd-Cd)Cb
L5: Cds-CddCbCt
L6: Cds-(Cdd-O2d)CbCt
L6: Cds-(Cdd-S2d)CbCt
L6: Cds-(Cdd-Cd)CbCt
L5: Cds-CddCbCb
L6: Cds-(Cdd-O2d)CbCb
L6: Cds-(Cdd-S2d)CbCb
L6: Cds-(Cdd-Cd)CbCb
L5: Cds-CdsC=SC=S
L5: Cds-(Cdd-Cd)C=S(Cds-Cd)
L6: Cds-(Cdd-Cd)C=S(Cds-Cds)
L6: Cds-(Cdd-Cd)C=S(Cds-Cdd)
L7: Cds-(Cdd-Cd)C=S(Cds-Cdd-Cd)
L7: Cds-(Cdd-Cd)C=S(Cds-Cdd-S2d)
L5: Cds-(Cdd-S2d)C=SCs
L5: Cds-(Cdd-S2d)C=SCt
L5: Cds-(Cdd-S2d)C=SCb
L5: Cds-(Cdd-Cd)C=SC=S
L5: Cds-(Cdd-S2d)(Cds-Cd)C=S
L6: Cds-(Cdd-S2d)(Cds-Cds)C=S
L6: Cds-(Cdd-S2d)(Cds-Cdd)C=S
L7: Cds-(Cdd-S2d)(Cds-Cdd-S2d)C=S
L7: Cds-(Cdd-S2d)(Cds-Cdd-Cd)C=S
L5: Cds-CdsCbC=S
L5: Cds-CdsCtC=S
L5: Cds-CdsC=SCs
L5: Cds-CdsC=S(Cds-Cd)
L6: Cds-CdsC=S(Cds-Cds)
L6: Cds-CdsC=S(Cds-Cdd)
L7: Cds-CdsC=S(Cds-Cdd-Cd)
L7: Cds-CdsC=S(Cds-Cdd-S2d)
L5: Cds-(Cdd-S2d)C=SC=S
L4: Cds-CNH
L5: Cd-CdHN3s
L5: Cd-CdH(N5dOdOs)
L4: Cds-CCN
L5: Cd-CdCsN3s
L5: Cd-CdCs(N5dOdOs)
L4: C=S-SsSs
L4: C=S-CH
L5: C=S-CsH
L5: C=S-CdsH
L6: C=S-(Cds-Cd)H
L7: C=S-(Cds-Cdd)H
L8: C=S-(Cds-Cdd-Cd)H
L8: C=S-(Cds-Cdd-S2d)H
L7: C=S-(Cds-Cds)H
L5: C=S-CtH
L5: C=S-CbH
L5: C=S-C=SH
L4: C=S-CC
L5: C=S-CbCds
L6: C=S-Cb(Cds-Cd)
L7: C=S-Cb(Cds-Cds)
L7: C=S-Cb(Cds-Cdd)
L8: C=S-Cb(Cds-Cdd-S2d)
L8: C=S-Cb(Cds-Cdd-Cd)
L5: C=S-CtCt
L5: C=S-CbCb
L5: C=S-CdsCds
L6: C=S-(Cds-Cd)(Cds-Cd)
L7: C=S-(Cds-Cdd)(Cds-Cds)
L8: C=S-(Cds-Cdd-Cd)(Cds-Cds)
L8: C=S-(Cds-Cdd-S2d)(Cds-Cds)
L7: C=S-(Cds-Cds)(Cds-Cds)
L7: C=S-(Cds-Cdd)(Cds-Cdd)
L8: C=S-(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L8: C=S-(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L8: C=S-(Cds-Cdd-Cd)(Cds-Cdd-S2d)
L5: C=S-CtCds
L6: C=S-Ct(Cds-Cd)
L7: C=S-Ct(Cds-Cds)
L7: C=S-Ct(Cds-Cdd)
L8: C=S-Ct(Cds-Cdd-Cd)
L8: C=S-Ct(Cds-Cdd-S2d)
L5: C=S-CbCt
L5: C=S-CsCs
L5: C=S-CdsCs
L6: C=S-(Cds-Cd)Cs
L7: C=S-(Cds-Cds)Cs
L7: C=S-(Cds-Cdd)Cs
L8: C=S-(Cds-Cdd-S2d)Cs
L8: C=S-(Cds-Cdd-Cd)Cs
L5: C=S-CtCs
L5: C=S-CbCs
L5: C=S-C=SCs
L5: C=S-CtC=S
L5: C=S-(Cds-Cd)C=S
L6: C=S-(Cds-Cdd)C=S
L7: C=S-(Cds-Cdd-Cd)C=S
L7: C=S-(Cds-Cdd-S2d)C=S
L6: C=S-(Cds-Cds)C=S
L5: C=S-C=SC=S
L5: C=S-CbC=S
L4: C=S-HH
L4: C=S-SsH
L4: C=S-CSs
L5: C=S-CbSs
L5: C=S-CdsSs
L6: C=S-(Cds-Cd)S2s
L7: C=S-(Cds-Cds)S2s
L7: C=S-(Cds-Cdd)S2s
L8: C=S-(Cds-Cdd-Cd)S2s
L8: C=S-(Cds-Cdd-S2d)S2s
L5: C=S-CtSs
L5: C=S-CsSs
L5: C=S-C=SSs
L4: CS-OsH
L4: CS-CsOs
L4: CS-OsOs
L3: Cs
L4: Cs-NHHH
L5: Cs-N3sHHH
L5: Cs-N3dHHH
L6: Cs-(N3dCd)HHH
L6: Cs-(N3dN3d)HHH
L4: Cs-NCsHH
L5: Cs-N3sCsHH
L5: Cs-N3dCHH
L6: Cs-(N3dN3d)CsHH
L6: Cs-(N3dOd)CHH
L6: Cs-(N3dCd)CsHH
L5: Cs-N5dCsHH
L6: Cs-(N5dOdOs)CsHH
L4: Cs-NCsCsH
L5: Cs-N3sCsCsH
L5: Cs-N3dCsCsH
L6: Cs-(N3dOd)CsCsH
L6: Cs-(N3dN3d)CsCsH
L5: Cs-N5dCsCsH
L6: Cs-(N5dOdOs)CsCsH
L4: Cs-NCsCsCs
L5: Cs-N3sCsCsCs
L5: Cs-N3dCsCsCs
L6: Cs-(N3dOd)CsCsCs
L6: Cs-(N3dN3d)CsCsCs
L5: Cs-N5dCsCsCs
L6: Cs-(N5dOdOs)CsCsCs
L4: Cs-NNCsCs
L5: Cs-N5dN5dCsCs
L4: Cs-NNCsH
L5: Cs-(N5dOdOs)(N5dOdOs)CsH
L4: Cs-HHHH
L4: Cs-CHHH
L5: Cs-CsHHH
L5: Cs-CdsHHH
L6: Cs-(Cds-O2d)HHH
L6: Cs-(Cds-Cd)HHH
L7: Cs-(Cds-Cds)HHH
L7: Cs-(Cds-Cdd)HHH
L8: Cs-(Cds-Cdd-O2d)HHH
L8: Cs-(Cds-Cdd-S2d)HHH
L8: Cs-(Cds-Cdd-Cd)HHH
L6: Cs-(CdN3d)HHH
L5: Cs-CtHHH
L6: Cs-(CtN3t)HHH
L5: Cs-CbHHH
L5: Cs-C=SHHH
L4: Cs-OsHHH
L4: Cs-OsOsHH
L4: Cs-OsOsOsH
L4: Cs-OsSsHH
L4: Cs-OsOsSsH
L4: Cs-SsHHH
L4: Cs-SsSsHH
L4: Cs-SsSsSsH
L4: Cs-CCHH
L5: Cs-CsCsHH
L5: Cs-CdsCsHH
L6: Cs-(Cds-O2d)CsHH
L6: Cs-(Cds-Cd)CsHH
L7: Cs-(Cds-Cds)CsHH
L7: Cs-(Cds-Cdd)CsHH
L8: Cs-(Cds-Cdd-O2d)CsHH
L8: Cs-(Cds-Cdd-S2d)CsHH
L8: Cs-(Cds-Cdd-Cd)CsHH
L6: Cs-(CdN3d)CsHH
L5: Cs-CdsCdsHH
L6: Cs-(Cds-O2d)(Cds-O2d)HH
L6: Cs-(Cds-O2d)(Cds-Cd)HH
L7: Cs-(Cds-O2d)(Cds-Cds)HH
L7: Cs-(Cds-O2d)(Cds-Cdd)HH
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)HH
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)HH
L6: Cs-(Cds-Cd)(Cds-Cd)HH
L7: Cs-(Cds-Cds)(Cds-Cds)HH
L7: Cs-(Cds-Cdd)(Cds-Cds)HH
L8: Cs-Cd(CCO)HH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)HH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)HH
L7: Cs-(Cds-Cdd)(Cds-Cdd)HH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)HH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)HH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)HH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)HH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)HH
L5: Cs-CtCsHH
L6: Cs-(CtN3t)CsHH
L5: Cs-CtCdsHH
L6: Cs-(Cds-O2d)CtHH
L6: Cs-(Cds-Cd)CtHH
L7: Cs-(Cds-Cds)CtHH
L7: Cs-(Cds-Cdd)CtHH
L8: Cs-(Cds-Cdd-O2d)CtHH
L8: Cs-(Cds-Cdd-S2d)CtHH
L8: Cs-(Cds-Cdd-Cd)CtHH
L5: Cs-CtCtHH
L5: Cs-CbCsHH
L5: Cs-CbCdsHH
L6: Cs-(Cds-O2d)CbHH
L6: Cs-(Cds-Cd)CbHH
L7: Cs-(Cds-Cds)CbHH
L7: Cs-(Cds-Cdd)CbHH
L8: Cs-(Cds-Cdd-O2d)CbHH
L8: Cs-(Cds-Cdd-S2d)CbHH
L8: Cs-(Cds-Cdd-Cd)CbHH
L5: Cs-CbCtHH
L5: Cs-CbCbHH
L5: Cs-C=SCtHH
L5: Cs-C=SCsHH
L5: Cs-C=S(Cds-Cd)HH
L6: Cs-C=S(Cds-Cdd)HH
L7: Cs-C=S(Cds-Cdd-Cd)HH
L7: Cs-C=S(Cds-Cdd-S2d)HH
L6: Cs-C=S(Cds-Cds)HH
L5: Cs-C=SC=SHH
L5: Cs-C=SCbHH
L4: Cs-CCCH
L5: Cs-CsCsCsH
L5: Cs-CdsCsCsH
L6: Cs-(Cds-O2d)CsCsH
L6: Cs-(Cds-Cd)CsCsH
L7: Cs-(Cds-Cds)CsCsH
L7: Cs-(Cds-Cdd)CsCsH
L8: Cs-(Cds-Cdd-O2d)CsCsH
L8: Cs-(Cds-Cdd-S2d)CsCsH
L8: Cs-(Cds-Cdd-Cd)CsCsH
L6: Cs-(CdN3d)CsCsH
L5: Cs-CtCsCsH
L6: Cs-(CtN3t)CsCsH
L5: Cs-CbCsCsH
L5: Cs-CdsCdsCsH
L6: Cs-(Cds-O2d)(Cds-O2d)CsH
L6: Cs-(Cds-O2d)(Cds-Cd)CsH
L7: Cs-(Cds-O2d)(Cds-Cds)CsH
L7: Cs-(Cds-O2d)(Cds-Cdd)CsH
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CsH
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CsH
L6: Cs-(Cds-Cd)(Cds-Cd)CsH
L7: Cs-(Cds-Cds)(Cds-Cds)CsH
L7: Cs-(Cds-Cdd)(Cds-Cds)CsH
L8: Cs-CsCd(CCO)H
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CsH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CsH
L7: Cs-(Cds-Cdd)(Cds-Cdd)CsH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsH
L5: Cs-CtCdsCsH
L6: Cs-(Cds-O2d)CtCsH
L6: Cs-(Cds-Cd)CtCsH
L7: Cs-(Cds-Cds)CtCsH
L7: Cs-(Cds-Cdd)CtCsH
L8: Cs-(Cds-Cdd-O2d)CtCsH
L8: Cs-(Cds-Cdd-S2d)CtCsH
L8: Cs-(Cds-Cdd-Cd)CtCsH
L5: Cs-CbCdsCsH
L6: Cs-(Cds-O2d)CbCsH
L6: Cs-(Cds-Cd)CbCsH
L7: Cs-(Cds-Cds)CbCsH
L7: Cs-(Cds-Cdd)CbCsH
L8: Cs-(Cds-Cdd-O2d)CbCsH
L8: Cs-(Cds-Cdd-Cd)CbCsH
L5: Cs-CtCtCsH
L5: Cs-CbCtCsH
L5: Cs-CbCbCsH
L5: Cs-CdsCdsCdsH
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)H
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)H
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)H
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)H
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)H
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)H
L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)H
L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)H
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)H
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)H
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)H
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)H
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H
L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)H
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)H
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)H
L8: Cs-CdCd(CCO)H
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)H
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)H
L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)H
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)H
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)H
L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H
L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)H
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)H
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)H
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)H
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)H
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)H
L5: Cs-CtCdsCdsH
L6: Cs-(Cds-O2d)(Cds-O2d)CtH
L6: Cs-(Cds-O2d)(Cds-Cd)CtH
L7: Cs-(Cds-O2d)(Cds-Cds)CtH
L7: Cs-(Cds-O2d)(Cds-Cdd)CtH
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CtH
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CtH
L6: Cs-(Cds-Cd)(Cds-Cd)CtH
L7: Cs-(Cds-Cds)(Cds-Cds)CtH
L7: Cs-(Cds-Cdd)(Cds-Cds)CtH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CtH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CtH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtH
L7: Cs-(Cds-Cdd)(Cds-Cdd)CtH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtH
L5: Cs-CbCdsCdsH
L6: Cs-(Cds-O2d)(Cds-O2d)CbH
L6: Cs-(Cds-O2d)(Cds-Cd)CbH
L7: Cs-(Cds-O2d)(Cds-Cds)CbH
L7: Cs-(Cds-O2d)(Cds-Cdd)CbH
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbH
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbH
L6: Cs-(Cds-Cd)(Cds-Cd)CbH
L7: Cs-(Cds-Cds)(Cds-Cds)CbH
L7: Cs-(Cds-Cdd)(Cds-Cds)CbH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbH
L7: Cs-(Cds-Cdd)(Cds-Cdd)CbH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbH
L5: Cs-CtCtCdsH
L6: Cs-CtCt(Cds-O2d)H
L6: Cs-CtCt(Cds-Cd)H
L7: Cs-CtCt(Cds-Cds)H
L7: Cs-CtCt(Cds-Cdd)H
L8: Cs-CtCt(Cds-Cdd-O2d)H
L8: Cs-CtCt(Cds-Cdd-S2d)H
L8: Cs-CtCt(Cds-Cdd-Cd)H
L5: Cs-CbCtCdsH
L6: Cs-CbCt(Cds-O2d)H
L6: Cs-CbCt(Cds-Cd)H
L7: Cs-CbCt(Cds-Cds)H
L7: Cs-CbCt(Cds-Cdd)H
L8: Cs-CbCt(Cds-Cdd-O2d)H
L8: Cs-CbCt(Cds-Cdd-S2d)H
L8: Cs-CbCt(Cds-Cdd-Cd)H
L5: Cs-CbCbCdsH
L6: Cs-CbCb(Cds-O2d)H
L6: Cs-CbCb(Cds-Cd)H
L7: Cs-CbCb(Cds-Cds)H
L7: Cs-CbCb(Cds-Cdd)H
L8: Cs-CbCb(Cds-Cdd-O2d)H
L8: Cs-CbCb(Cds-Cdd-S2d)H
L8: Cs-CbCb(Cds-Cdd-Cd)H
L5: Cs-CtCtCtH
L5: Cs-CbCtCtH
L5: Cs-CbCbCtH
L5: Cs-CbCbCbH
L5: Cs-C=SC=SCbH
L5: Cs-C=S(Cds-Cd)(Cds-Cd)H
L6: Cs-C=S(Cds-Cdd)(Cds-Cds)H
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)H
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)H
L6: Cs-C=S(Cds-Cds)(Cds-Cds)H
L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)H
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)H
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)H
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)H
L5: Cs-C=S(Cds-Cd)CtH
L6: Cs-C=S(Cds-Cdd)CtH
L7: Cs-C=S(Cds-Cdd-S2d)CtH
L7: Cs-C=S(Cds-Cdd-Cd)CtH
L6: Cs-C=S(Cds-Cds)CtH
L5: Cs-C=SC=SCtH
L5: Cs-C=SCtCsH
L5: Cs-C=SC=SCsH
L5: Cs-C=S(Cds-Cd)CbH
L6: Cs-C=S(Cds-Cds)CbH
L6: Cs-C=S(Cds-Cdd)CbH
L7: Cs-C=S(Cds-Cdd-S2d)CbH
L7: Cs-C=S(Cds-Cdd-Cd)CbH
L5: Cs-C=S(Cds-Cd)CsH
L6: Cs-C=S(Cds-Cds)CsH
L6: Cs-C=S(Cds-Cdd)CsH
L7: Cs-C=S(Cds-Cdd-Cd)CsH
L7: Cs-C=S(Cds-Cdd-S2d)CsH
L5: Cs-CbCtC=SH
L5: Cs-C=SC=SC=SH
L5: Cs-C=SCsCsH
L5: Cs-CtCtC=SH
L5: Cs-CbCbC=SH
L5: Cs-C=SC=S(Cds-Cd)H
L6: Cs-C=SC=S(Cds-Cds)H
L6: Cs-C=SC=S(Cds-Cdd)H
L7: Cs-C=SC=S(Cds-Cdd-S2d)H
L7: Cs-C=SC=S(Cds-Cdd-Cd)H
L4: Cs-CCCC
L5: Cs-CsCsCsCs
L5: Cs-CdsCsCsCs
L6: Cs-(Cds-O2d)CsCsCs
L6: Cs-(Cds-Cd)CsCsCs
L7: Cs-(Cds-Cds)CsCsCs
L7: Cs-(Cds-Cdd)CsCsCs
L8: Cs-(Cds-Cdd-O2d)CsCsCs
L8: Cs-(Cds-Cdd-S2d)CsCsCs
L8: Cs-(Cds-Cdd-Cd)CsCsCs
L6: Cs-(CdN3d)CsCsCs
L5: Cs-CtCsCsCs
L6: Cs-(CtN3t)CsCsCs
L5: Cs-CbCsCsCs
L5: Cs-CdsCdsCsCs
L6: Cs-(Cds-O2d)(Cds-O2d)CsCs
L6: Cs-(Cds-O2d)(Cds-Cd)CsCs
L7: Cs-(Cds-O2d)(Cds-Cds)CsCs
L7: Cs-(Cds-O2d)(Cds-Cdd)CsCs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CsCs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CsCs
L6: Cs-(Cds-Cd)(Cds-Cd)CsCs
L7: Cs-(Cds-Cds)(Cds-Cds)CsCs
L7: Cs-(Cds-Cdd)(Cds-Cds)CsCs
L8: Cs-CsCsCd(CCO)
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CsCs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CsCs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CsCs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsCs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsCs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsCs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsCs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsCs
L5: Cs-CtCdsCsCs
L6: Cs-(Cds-O2d)CtCsCs
L6: Cs-(Cds-Cd)CtCsCs
L7: Cs-(Cds-Cds)CtCsCs
L7: Cs-(Cds-Cdd)CtCsCs
L8: Cs-(Cds-Cdd-O2d)CtCsCs
L8: Cs-(Cds-Cdd-S2d)CtCsCs
L8: Cs-(Cds-Cdd-Cd)CtCsCs
L5: Cs-CbCdsCsCs
L6: Cs-(Cds-O2d)CbCsCs
L6: Cs-(Cds-Cd)CbCsCs
L7: Cs-(Cds-Cds)CbCsCs
L7: Cs-(Cds-Cdd)CbCsCs
L8: Cs-(Cds-Cdd-O2d)CbCsCs
L8: Cs-(Cds-Cdd-S2d)CbCsCs
L8: Cs-(Cds-Cdd-Cd)CbCsCs
L5: Cs-CtCtCsCs
L6: Cs-(CtN3t)(CtN3t)CsCs
L5: Cs-CbCtCsCs
L5: Cs-CbCbCsCs
L5: Cs-CdsCdsCdsCs
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cs
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Cs
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cs
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Cs
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cs
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cs
L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Cs
L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cs
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Cs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cs
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Cs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs
L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Cs
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cs
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cs
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cs
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Cs
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cs
L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Cs
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs
L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs
L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Cs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs
L5: Cs-CtCdsCdsCs
L6: Cs-(Cds-O2d)(Cds-O2d)CtCs
L6: Cs-(Cds-O2d)(Cds-Cd)CtCs
L7: Cs-(Cds-O2d)(Cds-Cds)CtCs
L7: Cs-(Cds-O2d)(Cds-Cdd)CtCs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CtCs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCs
L6: Cs-(Cds-Cd)(Cds-Cd)CtCs
L7: Cs-(Cds-Cds)(Cds-Cds)CtCs
L7: Cs-(Cds-Cdd)(Cds-Cds)CtCs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CtCs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CtCs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtCs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtCs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtCs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtCs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCs
L5: Cs-CbCdsCdsCs
L6: Cs-(Cds-O2d)(Cds-O2d)CbCs
L6: Cs-(Cds-O2d)(Cds-Cd)CbCs
L7: Cs-(Cds-O2d)(Cds-Cds)CbCs
L7: Cs-(Cds-O2d)(Cds-Cdd)CbCs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCs
L6: Cs-(Cds-Cd)(Cds-Cd)CbCs
L7: Cs-(Cds-Cds)(Cds-Cds)CbCs
L7: Cs-(Cds-Cdd)(Cds-Cds)CbCs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CbCs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCs
L5: Cs-CtCtCdsCs
L6: Cs-(Cds-O2d)CtCtCs
L6: Cs-(Cds-Cd)CtCtCs
L7: Cs-(Cds-Cds)CtCtCs
L7: Cs-(Cds-Cdd)CtCtCs
L8: Cs-(Cds-Cdd-O2d)CtCtCs
L8: Cs-(Cds-Cdd-S2d)CtCtCs
L8: Cs-(Cds-Cdd-Cd)CtCtCs
L5: Cs-CbCtCdsCs
L6: Cs-(Cds-O2d)CbCtCs
L6: Cs-(Cds-Cd)CbCtCs
L7: Cs-(Cds-Cds)CbCtCs
L7: Cs-(Cds-Cdd)CbCtCs
L8: Cs-(Cds-Cdd-O2d)CbCtCs
L8: Cs-(Cds-Cdd-S2d)CbCtCs
L8: Cs-(Cds-Cdd-Cd)CbCtCs
L5: Cs-CbCbCdsCs
L6: Cs-(Cds-O2d)CbCbCs
L6: Cs-(Cds-Cd)CbCbCs
L7: Cs-(Cds-Cds)CbCbCs
L7: Cs-(Cds-Cdd)CbCbCs
L8: Cs-(Cds-Cdd-O2d)CbCbCs
L8: Cs-(Cds-Cdd-S2d)CbCbCs
L8: Cs-(Cds-Cdd-Cd)CbCbCs
L5: Cs-CtCtCtCs
L5: Cs-CbCtCtCs
L5: Cs-CbCbCtCs
L5: Cs-CbCbCbCs
L5: Cs-CdsCdsCdsCds
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-O2d)
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cd)
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cds)
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd)
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)(Cds-Cd)
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)(Cds-Cds)
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)(Cds-Cds)
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)(Cds-Cdd)
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)(Cds-Cd)
L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cds)
L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd)
L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)
L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)
L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd)(Cds-Cdd)
L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)(Cds-Cd)
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cds)
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd)
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)(Cds-Cdd)
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L5: Cs-CtCdsCdsCds
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Ct
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Ct
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Ct
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Ct
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Ct
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Ct
L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Ct
L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Ct
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Ct
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Ct
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Ct
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Ct
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct
L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Ct
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Ct
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Ct
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Ct
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Ct
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Ct
L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Ct
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct
L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct
L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Ct
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Ct
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Ct
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct
L5: Cs-CbCdsCdsCds
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)Cb
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)Cb
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)Cb
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)Cb
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)Cb
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)Cb
L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)Cb
L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)Cb
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)Cb
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)Cb
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)Cb
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)Cb
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb
L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)Cb
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)Cb
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)Cb
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)Cb
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)Cb
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)Cb
L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)Cb
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb
L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb
L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)Cb
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)Cb
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)Cb
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb
L5: Cs-CtCtCdsCds
L6: Cs-(Cds-O2d)(Cds-O2d)CtCt
L6: Cs-(Cds-O2d)(Cds-Cd)CtCt
L7: Cs-(Cds-O2d)(Cds-Cds)CtCt
L7: Cs-(Cds-O2d)(Cds-Cdd)CtCt
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CtCt
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CtCt
L6: Cs-(Cds-Cd)(Cds-Cd)CtCt
L7: Cs-(Cds-Cds)(Cds-Cds)CtCt
L7: Cs-(Cds-Cdd)(Cds-Cds)CtCt
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CtCt
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CtCt
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtCt
L7: Cs-(Cds-Cdd)(Cds-Cdd)CtCt
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtCt
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtCt
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtCt
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtCt
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtCt
L5: Cs-CbCtCdsCds
L6: Cs-(Cds-O2d)(Cds-O2d)CbCt
L6: Cs-(Cds-O2d)(Cds-Cd)CbCt
L7: Cs-(Cds-O2d)(Cds-Cds)CbCt
L7: Cs-(Cds-O2d)(Cds-Cdd)CbCt
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCt
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCt
L6: Cs-(Cds-Cd)(Cds-Cd)CbCt
L7: Cs-(Cds-Cds)(Cds-Cds)CbCt
L7: Cs-(Cds-Cdd)(Cds-Cds)CbCt
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCt
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCt
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCt
L7: Cs-(Cds-Cdd)(Cds-Cdd)CbCt
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCt
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCt
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCt
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCt
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCt
L5: Cs-CbCbCdsCds
L6: Cs-(Cds-O2d)(Cds-O2d)CbCb
L6: Cs-(Cds-O2d)(Cds-Cd)CbCb
L7: Cs-(Cds-O2d)(Cds-Cds)CbCb
L7: Cs-(Cds-O2d)(Cds-Cdd)CbCb
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbCb
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbCb
L6: Cs-(Cds-Cd)(Cds-Cd)CbCb
L7: Cs-(Cds-Cds)(Cds-Cds)CbCb
L7: Cs-(Cds-Cdd)(Cds-Cds)CbCb
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbCb
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbCb
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbCb
L7: Cs-(Cds-Cdd)(Cds-Cdd)CbCb
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbCb
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbCb
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbCb
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbCb
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbCb
L5: Cs-CtCtCtCds
L6: Cs-(Cds-O2d)CtCtCt
L6: Cs-(Cds-Cd)CtCtCt
L7: Cs-(Cds-Cds)CtCtCt
L7: Cs-(Cds-Cdd)CtCtCt
L8: Cs-(Cds-Cdd-O2d)CtCtCt
L8: Cs-(Cds-Cdd-S2d)CtCtCt
L8: Cs-(Cds-Cdd-Cd)CtCtCt
L5: Cs-CbCtCtCds
L6: Cs-(Cds-O2d)CbCtCt
L6: Cs-(Cds-Cd)CbCtCt
L7: Cs-(Cds-Cds)CbCtCt
L7: Cs-(Cds-Cdd)CbCtCt
L8: Cs-(Cds-Cdd-O2d)CbCtCt
L8: Cs-(Cds-Cdd-S2d)CbCtCt
L8: Cs-(Cds-Cdd-Cd)CbCtCt
L5: Cs-CbCbCtCds
L6: Cs-(Cds-O2d)CbCbCt
L6: Cs-(Cds-Cd)CbCbCt
L7: Cs-(Cds-Cds)CbCbCt
L7: Cs-(Cds-Cdd)CbCbCt
L8: Cs-(Cds-Cdd-O2d)CbCbCt
L8: Cs-(Cds-Cdd-S2d)CbCbCt
L8: Cs-(Cds-Cdd-Cd)CbCbCt
L5: Cs-CbCbCbCds
L6: Cs-(Cds-O2d)CbCbCb
L6: Cs-(Cds-Cd)CbCbCb
L7: Cs-(Cds-Cds)CbCbCb
L7: Cs-(Cds-Cdd)CbCbCb
L8: Cs-(Cds-Cdd-O2d)CbCbCb
L8: Cs-(Cds-Cdd-S2d)CbCbCb
L8: Cs-(Cds-Cdd-Cd)CbCbCb
L5: Cs-CtCtCtCt
L5: Cs-CbCtCtCt
L5: Cs-CbCbCtCt
L5: Cs-CbCbCbCt
L5: Cs-CbCbCbCb
L5: Cs-C=SCbCtCt
L5: Cs-C=S(Cds-Cd)(Cds-Cd)(Cds-Cd)
L6: Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd)
L7: Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)
L7: Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)
L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L6: Cs-C=S(Cds-Cds)(Cds-Cds)(Cds-Cds)
L6: Cs-C=S(Cds-Cds)(Cds-Cdd)(Cds-Cdd)
L7: Cs-C=S(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L7: Cs-C=S(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L7: Cs-C=S(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L5: Cs-C=S(Cds-Cd)CtCt
L6: Cs-C=S(Cds-Cds)CtCt
L6: Cs-C=S(Cds-Cdd)CtCt
L7: Cs-C=S(Cds-Cdd-S2d)CtCt
L7: Cs-C=S(Cds-Cdd-Cd)CtCt
L5: Cs-C=S(Cds-Cd)CtCs
L6: Cs-C=S(Cds-Cds)CtCs
L6: Cs-C=S(Cds-Cdd)CtCs
L7: Cs-C=S(Cds-Cdd-S2d)CtCs
L7: Cs-C=S(Cds-Cdd-Cd)CtCs
L5: Cs-C=SCbCbCt
L5: Cs-C=SCbCsCs
L5: Cs-C=SCbCbCs
L5: Cs-C=SCtCtCt
L5: Cs-C=S(Cds-Cd)(Cds-Cd)Cs
L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)Cs
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cs
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cs
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cs
L6: Cs-C=S(Cds-Cds)(Cds-Cds)Cs
L6: Cs-C=S(Cds-Cdd)(Cds-Cds)Cs
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Cs
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Cs
L5: Cs-C=SC=SCtCt
L5: Cs-C=SCsCsCs
L5: Cs-C=SCtCtCs
L5: Cs-C=SC=SC=SCt
L5: Cs-C=SC=SC=SCs
L5: Cs-C=SC=SC=SC=S
L5: Cs-C=SCtCsCs
L5: Cs-C=SC=SC=SCb
L5: Cs-C=SC=SC=S(Cds-Cd)
L6: Cs-C=SC=SC=S(Cds-Cdd)
L7: Cs-C=SC=SC=S(Cds-Cdd-Cd)
L7: Cs-C=SC=SC=S(Cds-Cdd-S2d)
L6: Cs-C=SC=SC=S(Cds-Cds)
L5: Cs-C=S(Cds-Cd)(Cds-Cd)Ct
L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)Ct
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Ct
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Ct
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Ct
L6: Cs-C=S(Cds-Cds)(Cds-Cds)Ct
L6: Cs-C=S(Cds-Cdd)(Cds-Cds)Ct
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Ct
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Ct
L5: Cs-C=SC=SCtCs
L5: Cs-C=SC=SCbCb
L5: Cs-C=S(Cds-Cd)CsCs
L6: Cs-C=S(Cds-Cds)CsCs
L6: Cs-C=S(Cds-Cdd)CsCs
L7: Cs-C=S(Cds-Cdd-Cd)CsCs
L7: Cs-C=S(Cds-Cdd-S2d)CsCs
L5: Cs-C=SC=SCbCt
L5: Cs-C=S(Cds-Cd)CbCt
L6: Cs-C=S(Cds-Cds)CbCt
L6: Cs-C=S(Cds-Cdd)CbCt
L7: Cs-C=S(Cds-Cdd-S2d)CbCt
L7: Cs-C=S(Cds-Cdd-Cd)CbCt
L5: Cs-C=SC=SCsCs
L5: Cs-C=S(Cds-Cd)CbCb
L6: Cs-C=S(Cds-Cds)CbCb
L6: Cs-C=S(Cds-Cdd)CbCb
L7: Cs-C=S(Cds-Cdd-S2d)CbCb
L7: Cs-C=S(Cds-Cdd-Cd)CbCb
L5: Cs-C=SC=S(Cds-Cd)Ct
L6: Cs-C=SC=S(Cds-Cds)Ct
L6: Cs-C=SC=S(Cds-Cdd)Ct
L7: Cs-C=SC=S(Cds-Cdd-Cd)Ct
L7: Cs-C=SC=S(Cds-Cdd-S2d)Ct
L5: Cs-C=SC=S(Cds-Cd)Cs
L6: Cs-C=SC=S(Cds-Cds)Cs
L6: Cs-C=SC=S(Cds-Cdd)Cs
L7: Cs-C=SC=S(Cds-Cdd-S2d)Cs
L7: Cs-C=SC=S(Cds-Cdd-Cd)Cs
L5: Cs-C=SC=S(Cds-Cd)(Cds-Cd)
L6: Cs-C=SC=S(Cds-Cdd)(Cds-Cds)
L7: Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cds)
L7: Cs-C=SC=S(Cds-Cdd-Cd)(Cds-Cds)
L6: Cs-C=SC=S(Cds-Cdd)(Cds-Cdd)
L7: Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)
L7: Cs-C=SC=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)
L7: Cs-C=SC=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)
L6: Cs-C=SC=S(Cds-Cds)(Cds-Cds)
L5: Cs-C=SC=S(Cds-Cd)Cb
L6: Cs-C=SC=S(Cds-Cdd)Cb
L7: Cs-C=SC=S(Cds-Cdd-S2d)Cb
L7: Cs-C=SC=S(Cds-Cdd-Cd)Cb
L6: Cs-C=SC=S(Cds-Cds)Cb
L5: Cs-C=SCbCtCs
L5: Cs-C=S(Cds-Cd)CbCs
L6: Cs-C=S(Cds-Cds)CbCs
L6: Cs-C=S(Cds-Cdd)CbCs
L7: Cs-C=S(Cds-Cdd-S2d)CbCs
L7: Cs-C=S(Cds-Cdd-Cd)CbCs
L5: Cs-C=S(Cds-Cd)(Cds-Cd)Cb
L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)Cb
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)Cb
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)Cb
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)Cb
L6: Cs-C=S(Cds-Cds)(Cds-Cds)Cb
L6: Cs-C=S(Cds-Cdd)(Cds-Cds)Cb
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)Cb
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)Cb
L5: Cs-C=SCbCbCb
L5: Cs-C=SC=SCbCs
L4: Cs-CCCOs
L5: Cs-CsCsCsOs
L5: Cs-CdsCsCsOs
L6: Cs-(Cds-O2d)CsCsOs
L6: Cs-(Cds-Cd)CsCsOs
L7: Cs-(Cds-Cds)CsCsOs
L7: Cs-(Cds-Cdd)CsCsOs
L8: Cs-(Cds-Cdd-O2d)CsCsOs
L8: Cs-(Cds-Cdd-Cd)CsCsOs
L5: Cs-OsCtCsCs
L5: Cs-CbCsCsOs
L5: Cs-CdsCdsCsOs
L6: Cs-(Cds-O2d)(Cds-O2d)CsOs
L6: Cs-(Cds-O2d)(Cds-Cd)CsOs
L7: Cs-(Cds-O2d)(Cds-Cds)CsOs
L7: Cs-(Cds-O2d)(Cds-Cdd)CsOs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CsOs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CsOs
L6: Cs-(Cds-Cd)(Cds-Cd)CsOs
L7: Cs-(Cds-Cds)(Cds-Cds)CsOs
L7: Cs-(Cds-Cdd)(Cds-Cds)CsOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CsOs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CsOs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CsOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CsOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CsOs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsOs
L5: Cs-CtCdsCsOs
L6: Cs-(Cds-O2d)CtCsOs
L6: Cs-(Cds-Cd)CtCsOs
L7: Cs-(Cds-Cds)CtCsOs
L7: Cs-(Cds-Cdd)CtCsOs
L8: Cs-(Cds-Cdd-O2d)CtCsOs
L8: Cs-(Cds-Cdd-Cd)CtCsOs
L5: Cs-CbCdsCsOs
L6: Cs-(Cds-O2d)CbCsOs
L6: Cs-(Cds-Cd)CbCsOs
L7: Cs-(Cds-Cds)CbCsOs
L7: Cs-(Cds-Cdd)CbCsOs
L8: Cs-(Cds-Cdd-O2d)CbCsOs
L8: Cs-(Cds-Cdd-Cd)CbCsOs
L5: Cs-CtCtCsOs
L5: Cs-CbCtCsOs
L5: Cs-CbCbCsOs
L5: Cs-CdsCdsCdsOs
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-O2d)O2s
L6: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cd)O2s
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cds)O2s
L7: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd)O2s
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-O2d)O2s
L8: Cs-(Cds-O2d)(Cds-O2d)(Cds-Cdd-Cd)O2s
L6: Cs-(Cds-O2d)(Cds-Cd)(Cds-Cd)O2s
L7: Cs-(Cds-O2d)(Cds-Cds)(Cds-Cds)O2s
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cds)O2s
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cds)O2s
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cds)O2s
L7: Cs-(Cds-O2d)(Cds-Cdd)(Cds-Cdd)O2s
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s
L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)O2s
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)O2s
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)O2s
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-O2d)O2s
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)O2s
L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)O2s
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s
L8: Cs-(Cds-Cds)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s
L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s
L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)O2s
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-O2d)O2s
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)(Cds-Cdd-Cd)O2s
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)O2s
L5: Cs-CtCdsCdsOs
L6: Cs-(Cds-O2d)(Cds-O2d)CtOs
L6: Cs-(Cds-O2d)(Cds-Cd)CtOs
L7: Cs-(Cds-O2d)(Cds-Cds)CtOs
L7: Cs-(Cds-O2d)(Cds-Cdd)CtOs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CtOs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CtOs
L6: Cs-(Cds-Cd)(Cds-Cd)CtOs
L7: Cs-(Cds-Cds)(Cds-Cds)CtOs
L7: Cs-(Cds-Cdd)(Cds-Cds)CtOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CtOs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtOs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CtOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CtOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CtOs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtOs
L5: Cs-CbCdsCdsOs
L6: Cs-(Cds-O2d)(Cds-O2d)CbOs
L6: Cs-(Cds-O2d)(Cds-Cd)CbOs
L7: Cs-(Cds-O2d)(Cds-Cds)CbOs
L7: Cs-(Cds-O2d)(Cds-Cdd)CbOs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)CbOs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)CbOs
L6: Cs-(Cds-Cd)(Cds-Cd)CbOs
L7: Cs-(Cds-Cds)(Cds-Cds)CbOs
L7: Cs-(Cds-Cdd)(Cds-Cds)CbOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)CbOs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbOs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CbOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)CbOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)CbOs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbOs
L5: Cs-CtCtCdsOs
L6: Cs-(Cds-O2d)CtCtOs
L6: Cs-(Cds-Cd)CtCtOs
L7: Cs-(Cds-Cds)CtCtOs
L7: Cs-(Cds-Cdd)CtCtOs
L8: Cs-(Cds-Cdd-O2d)CtCtOs
L8: Cs-(Cds-Cdd-Cd)CtCtOs
L5: Cs-CbCtCdsOs
L6: Cs-(Cds-O2d)CbCtOs
L6: Cs-(Cds-Cd)CbCtOs
L7: Cs-(Cds-Cds)CbCtOs
L7: Cs-(Cds-Cdd)CbCtOs
L8: Cs-(Cds-Cdd-O2d)CbCtOs
L8: Cs-(Cds-Cdd-Cd)CbCtOs
L5: Cs-CbCbCdsOs
L6: Cs-(Cds-O2d)CbCbOs
L6: Cs-(Cds-Cd)CbCbOs
L7: Cs-(Cds-Cds)CbCbOs
L7: Cs-(Cds-Cdd)CbCbOs
L8: Cs-(Cds-Cdd-O2d)CbCbOs
L8: Cs-(Cds-Cdd-Cd)CbCbOs
L5: Cs-CtCtCtOs
L5: Cs-CbCtCtOs
L5: Cs-CbCbCtOs
L5: Cs-CbCbCbOs
L4: Cs-CCOsOs
L5: Cs-CsCsOsOs
L5: Cs-CdsCsOsOs
L6: Cs-(Cds-O2d)CsOsOs
L6: Cs-(Cds-Cd)CsOsOs
L7: Cs-(Cds-Cds)CsOsOs
L7: Cs-(Cds-Cdd)CsOsOs
L8: Cs-(Cds-Cdd-O2d)CsOsOs
L8: Cs-(Cds-Cdd-Cd)CsOsOs
L5: Cs-CdsCdsOsOs
L6: Cs-(Cds-O2d)(Cds-O2d)OsOs
L6: Cs-(Cds-O2d)(Cds-Cd)OsOs
L7: Cs-(Cds-O2d)(Cds-Cds)OsOs
L7: Cs-(Cds-O2d)(Cds-Cdd)OsOs
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)OsOs
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)OsOs
L6: Cs-(Cds-Cd)(Cds-Cd)OsOs
L7: Cs-(Cds-Cds)(Cds-Cds)OsOs
L7: Cs-(Cds-Cdd)(Cds-Cds)OsOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)OsOs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)OsOs
L7: Cs-(Cds-Cdd)(Cds-Cdd)OsOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)OsOs
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)OsOs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsOs
L5: Cs-CtCsOsOs
L5: Cs-CtCdsOsOs
L6: Cs-(Cds-O2d)CtOsOs
L6: Cs-(Cds-Cd)CtOsOs
L7: Cs-(Cds-Cds)CtOsOs
L7: Cs-(Cds-Cdd)CtOsOs
L8: Cs-(Cds-Cdd-O2d)CtOsOs
L8: Cs-(Cds-Cdd-Cd)CtOsOs
L5: Cs-CtCtOsOs
L5: Cs-CbCsOsOs
L5: Cs-CbCdsOsOs
L6: Cs-(Cds-O2d)CbOsOs
L6: Cs-(Cds-Cd)CbOsOs
L7: Cs-(Cds-Cds)CbOsOs
L7: Cs-(Cds-Cdd)CbOsOs
L8: Cs-(Cds-Cdd-O2d)CbOsOs
L8: Cs-(Cds-Cdd-Cd)CbOsOs
L5: Cs-CbCtOsOs
L5: Cs-CbCbOsOs
L4: Cs-COsOsOs
L5: Cs-CsOsOsOs
L5: Cs-CdsOsOsOs
L6: Cs-(Cds-O2d)OsOsOs
L6: Cs-(Cds-Cd)OsOsOs
L7: Cs-(Cds-Cds)OsOsOs
L7: Cs-(Cds-Cdd)OsOsOs
L8: Cs-(Cds-Cdd-O2d)OsOsOs
L8: Cs-(Cds-Cdd-Cd)OsOsOs
L5: Cs-CtOsOsOs
L5: Cs-CbOsOsOs
L4: Cs-OsOsOsOs
L4: Cs-COsOsH
L5: Cs-CsOsOsH
L5: Cs-CdsOsOsH
L6: Cs-(Cds-O2d)OsOsH
L6: Cs-(Cds-Cd)OsOsH
L7: Cs-(Cds-Cds)OsOsH
L7: Cs-(Cds-Cdd)OsOsH
L8: Cs-(Cds-Cdd-O2d)OsOsH
L8: Cs-(Cds-Cdd-Cd)OsOsH
L5: Cs-CtOsOsH
L5: Cs-CbOsOsH
L4: Cs-COsSsH
L5: Cs-CsOsSsH
L5: Cs-CdsOsSsH
L5: Cs-CtOsSsH
L5: Cs-CbOsSsH
L4: Cs-CCOsSs
L5: Cs-CsCsOsSs
L4: Cs-COsOsSs
L5: Cs-CsOsOsSs
L4: Cs-CCOsH
L5: Cs-CsCsOsH
L5: Cs-CdsCsOsH
L6: Cs-(Cds-O2d)CsOsH
L6: Cs-(Cds-Cd)CsOsH
L7: Cs-(Cds-Cds)CsOsH
L7: Cs-(Cds-Cdd)CsOsH
L8: Cs-(Cds-Cdd-O2d)CsOsH
L8: Cs-(Cds-Cdd-Cd)CsOsH
L5: Cs-CdsCdsOsH
L6: Cs-(Cds-O2d)(Cds-O2d)OsH
L6: Cs-(Cds-O2d)(Cds-Cd)OsH
L7: Cs-(Cds-O2d)(Cds-Cds)OsH
L7: Cs-(Cds-O2d)(Cds-Cdd)OsH
L8: Cs-(Cds-O2d)(Cds-Cdd-O2d)OsH
L8: Cs-(Cds-O2d)(Cds-Cdd-Cd)OsH
L6: Cs-(Cds-Cd)(Cds-Cd)OsH
L7: Cs-(Cds-Cds)(Cds-Cds)OsH
L7: Cs-(Cds-Cdd)(Cds-Cds)OsH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cds)OsH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)OsH
L7: Cs-(Cds-Cdd)(Cds-Cdd)OsH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-O2d)OsH
L8: Cs-(Cds-Cdd-O2d)(Cds-Cdd-Cd)OsH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)OsH
L5: Cs-CtCsOsH
L5: Cs-CtCdsOsH
L6: Cs-(Cds-O2d)CtOsH
L6: Cs-(Cds-Cd)CtOsH
L7: Cs-(Cds-Cds)CtOsH
L7: Cs-(Cds-Cdd)CtOsH
L8: Cs-(Cds-Cdd-O2d)CtOsH
L8: Cs-(Cds-Cdd-Cd)CtOsH
L5: Cs-CtCtOsH
L5: Cs-CbCsOsH
L5: Cs-CbCdsOsH
L6: Cs-(Cds-O2d)CbOsH
L6: Cs-(Cds-Cd)CbOsH
L7: Cs-(Cds-Cds)CbOsH
L7: Cs-(Cds-Cdd)CbOsH
L8: Cs-(Cds-Cdd-O2d)CbOsH
L8: Cs-(Cds-Cdd-Cd)CbOsH
L5: Cs-CbCtOsH
L5: Cs-CbCbOsH
L4: Cs-COsHH
L5: Cs-CsOsHH
L5: Cs-CdsOsHH
L6: Cs-(Cds-O2d)OsHH
L6: Cs-(Cds-Cd)OsHH
L7: Cs-(Cds-Cds)OsHH
L7: Cs-(Cds-Cdd)OsHH
L8: Cs-(Cds-Cdd-O2d)OsHH
L8: Cs-(Cds-Cdd-Cd)OsHH
L5: Cs-CtOsHH
L5: Cs-CbOsHH
L4: Cs-CCCSs
L5: Cs-CsCsCsSs
L5: Cs-CdsCsCsSs
L6: Cs-(Cds-Cd)CsCsSs
L7: Cs-(Cds-Cds)CsCsSs
L7: Cs-(Cds-Cdd)CsCsSs
L8: Cs-(Cds-Cdd-S2d)CsCsSs
L8: Cs-(Cds-Cdd-Cd)CsCsSs
L5: Cs-SsCtCsCs
L5: Cs-CbCsCsSs
L5: Cs-CdsCdsCsSs
L6: Cs-(Cds-Cd)(Cds-Cd)CsSs
L7: Cs-(Cds-Cds)(Cds-Cds)CsSs
L7: Cs-(Cds-Cdd)(Cds-Cds)CsSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CsSs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CsSs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CsSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CsSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CsSs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CsSs
L5: Cs-CtCdsCsSs
L6: Cs-(Cds-Cd)CtCsSs
L7: Cs-(Cds-Cds)CtCsSs
L7: Cs-(Cds-Cdd)CtCsSs
L8: Cs-(Cds-Cdd-S2d)CtCsSs
L8: Cs-(Cds-Cdd-Cd)CtCsSs
L5: Cs-CbCdsCsSs
L6: Cs-(Cds-Cd)CbCsSs
L7: Cs-(Cds-Cds)CbCsSs
L7: Cs-(Cds-Cdd)CbCsSs
L8: Cs-(Cds-Cdd-S2d)CbCsSs
L8: Cs-(Cds-Cdd-Cd)CbCsSs
L5: Cs-CtCtCsSs
L5: Cs-CbCtCsSs
L5: Cs-CbCbCsSs
L5: Cs-CdsCdsCdsSs
L6: Cs-(Cds-Cd)(Cds-Cd)(Cds-Cd)S2s
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cds)S2s
L7: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd)S2s
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-S2d)S2s
L8: Cs-(Cds-Cds)(Cds-Cds)(Cds-Cdd-Cd)S2s
L7: Cs-(Cds-Cds)(Cds-Cdd)(Cds-Cdd)S2s
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s
L8: Cs-(Cds-Cds)(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s
L8: Cs-(Cds-Cds)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s
L7: Cs-(Cds-Cdd)(Cds-Cdd)(Cds-Cdd)S2s
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s
L5: Cs-CtCdsCdsSs
L6: Cs-(Cds-Cd)(Cds-Cd)CtSs
L7: Cs-(Cds-Cds)(Cds-Cds)CtSs
L7: Cs-(Cds-Cdd)(Cds-Cds)CtSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CtSs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CtSs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CtSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CtSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CtSs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CtSs
L5: Cs-CbCdsCdsSs
L6: Cs-(Cds-Cd)(Cds-Cd)CbSs
L7: Cs-(Cds-Cds)(Cds-Cds)CbSs
L7: Cs-(Cds-Cdd)(Cds-Cds)CbSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)CbSs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)CbSs
L7: Cs-(Cds-Cdd)(Cds-Cdd)CbSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)CbSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)CbSs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)CbSs
L5: Cs-CtCtCdsSs
L6: Cs-(Cds-Cd)CtCtSs
L7: Cs-(Cds-Cds)CtCtSs
L7: Cs-(Cds-Cdd)CtCtSs
L8: Cs-(Cds-Cdd-S2d)CtCtSs
L8: Cs-(Cds-Cdd-Cd)CtCtSs
L5: Cs-CbCtCdsSs
L6: Cs-(Cds-Cd)CbCtSs
L7: Cs-(Cds-Cds)CbCtSs
L7: Cs-(Cds-Cdd)CbCtSs
L8: Cs-(Cds-Cdd-S2d)CbCtSs
L8: Cs-(Cds-Cdd-Cd)CbCtSs
L5: Cs-CbCbCdsSs
L6: Cs-(Cds-Cd)CbCbSs
L7: Cs-(Cds-Cds)CbCbSs
L7: Cs-(Cds-Cdd)CbCbSs
L8: Cs-(Cds-Cdd-S2d)CbCbSs
L8: Cs-(Cds-Cdd-Cd)CbCbSs
L5: Cs-CtCtCtSs
L5: Cs-CbCtCtSs
L5: Cs-CbCbCtSs
L5: Cs-CbCbCbSs
L5: Cs-C=SCbCsSs
L5: Cs-C=SCsCsSs
L5: Cs-C=S(Cds-Cd)(Cds-Cd)S2s
L6: Cs-C=S(Cds-Cdd)(Cds-Cdd)S2s
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cdd-Cd)S2s
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-Cd)S2s
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cdd-S2d)S2s
L6: Cs-C=S(Cds-Cdd)(Cds-Cds)S2s
L7: Cs-C=S(Cds-Cdd-Cd)(Cds-Cds)S2s
L7: Cs-C=S(Cds-Cdd-S2d)(Cds-Cds)S2s
L6: Cs-C=S(Cds-Cds)(Cds-Cds)S2s
L5: Cs-C=S(Cds-Cd)CtSs
L6: Cs-C=S(Cds-Cds)CtSs
L6: Cs-C=S(Cds-Cdd)CtSs
L7: Cs-C=S(Cds-Cdd-S2d)CtSs
L7: Cs-C=S(Cds-Cdd-Cd)CtSs
L5: Cs-C=SCtCsSs
L5: Cs-C=SC=SC=SSs
L5: Cs-C=SC=S(Cds-Cd)S2s
L6: Cs-C=SC=S(Cds-Cds)S2s
L6: Cs-C=SC=S(Cds-Cdd)S2s
L7: Cs-C=SC=S(Cds-Cdd-S2d)S2s
L7: Cs-C=SC=S(Cds-Cdd-Cd)S2s
L5: Cs-C=SCbCbSs
L5: Cs-C=SC=SCbSs
L5: Cs-C=SC=SCsSs
L5: Cs-C=SCtCtSs
L5: Cs-C=S(Cds-Cd)CbSs
L6: Cs-C=S(Cds-Cdd)CbSs
L7: Cs-C=S(Cds-Cdd-Cd)CbSs
L7: Cs-C=S(Cds-Cdd-S2d)CbSs
L6: Cs-C=S(Cds-Cds)CbSs
L5: Cs-C=SCbCtSs
L5: Cs-C=SC=SCtSs
L5: Cs-C=S(Cds-Cd)CsSs
L6: Cs-C=S(Cds-Cds)CsSs
L6: Cs-C=S(Cds-Cdd)CsSs
L7: Cs-C=S(Cds-Cdd-S2d)CsSs
L7: Cs-C=S(Cds-Cdd-Cd)CsSs
L4: Cs-CCSsSs
L5: Cs-CsCsSsSs
L5: Cs-CdsCsSsSs
L6: Cs-(Cds-Cd)CsSsSs
L7: Cs-(Cds-Cds)CsSsSs
L7: Cs-(Cds-Cdd)CsSsSs
L8: Cs-(Cds-Cdd-S2d)CsSsSs
L8: Cs-(Cds-Cdd-Cd)CsSsSs
L5: Cs-CdsCdsSsSs
L6: Cs-(Cds-Cd)(Cds-Cd)SsSs
L7: Cs-(Cds-Cds)(Cds-Cds)SsSs
L7: Cs-(Cds-Cdd)(Cds-Cds)SsSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)SsSs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)SsSs
L7: Cs-(Cds-Cdd)(Cds-Cdd)SsSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)SsSs
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)SsSs
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)SsSs
L5: Cs-CtCsSsSs
L5: Cs-CtCdsSsSs
L6: Cs-(Cds-Cd)CtSsSs
L7: Cs-(Cds-Cds)CtSsSs
L7: Cs-(Cds-Cdd)CtSsSs
L8: Cs-(Cds-Cdd-S2d)CtSsSs
L8: Cs-(Cds-Cdd-Cd)CtSsSs
L5: Cs-CtCtSsSs
L5: Cs-CbCsSsSs
L5: Cs-CbCdsSsSs
L6: Cs-(Cds-Cd)CbSsSs
L7: Cs-(Cds-Cds)CbSsSs
L7: Cs-(Cds-Cdd)CbSsSs
L8: Cs-(Cds-Cdd-S2d)CbSsSs
L8: Cs-(Cds-Cdd-Cd)CbSsSs
L5: Cs-CbCtSsSs
L5: Cs-CbCbSsSs
L5: Cs-C=SCsSsSs
L5: Cs-C=S(Cds-Cd)SsSs
L6: Cs-C=S(Cds-Cdd)SsSs
L7: Cs-C=S(Cds-Cdd-Cd)SsSs
L7: Cs-C=S(Cds-Cdd-S2d)SsSs
L6: Cs-C=S(Cds-Cds)SsSs
L5: Cs-C=SC=SSsSs
L5: Cs-C=SCbSsSs
L5: Cs-C=SCtSsSs
L4: Cs-CSsSsSs
L5: Cs-CsSsSsSs
L5: Cs-CdsSsSsSs
L6: Cs-(Cds-Cd)SsSsSs
L7: Cs-(Cds-Cds)SsSsSs
L7: Cs-(Cds-Cdd)SsSsSs
L8: Cs-(Cds-Cdd-S2d)SsSsSs
L8: Cs-(Cds-Cdd-Cd)SsSsSs
L5: Cs-CtSsSsSs
L5: Cs-CbSsSsSs
L5: Cs-C=SSsSsSs
L4: Cs-SsSsSsSs
L4: Cs-CSsSsH
L5: Cs-CsSsSsH
L5: Cs-CdsSsSsH
L6: Cs-(Cds-Cd)SsSsH
L7: Cs-(Cds-Cds)SsSsH
L7: Cs-(Cds-Cdd)SsSsH
L8: Cs-(Cds-Cdd-S2d)SsSsH
L8: Cs-(Cds-Cdd-Cd)SsSsH
L5: Cs-CtSsSsH
L5: Cs-CbSsSsH
L5: Cs-C=SSsSsH
L4: Cs-CCSsH
L5: Cs-CsCsSsH
L5: Cs-CdsCsSsH
L6: Cs-(Cds-Cd)CsSsH
L7: Cs-(Cds-Cds)CsSsH
L7: Cs-(Cds-Cdd)CsSsH
L8: Cs-(Cds-Cdd-S2d)CsSsH
L8: Cs-(Cds-Cdd-Cd)CsSsH
L5: Cs-CdsCdsSsH
L6: Cs-(Cds-Cd)(Cds-Cd)SsH
L7: Cs-(Cds-Cds)(Cds-Cds)SsH
L7: Cs-(Cds-Cdd)(Cds-Cds)SsH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cds)SsH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cds)SsH
L7: Cs-(Cds-Cdd)(Cds-Cdd)SsH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-S2d)SsH
L8: Cs-(Cds-Cdd-S2d)(Cds-Cdd-Cd)SsH
L8: Cs-(Cds-Cdd-Cd)(Cds-Cdd-Cd)SsH
L5: Cs-CtCsSsH
L5: Cs-CtCdsSsH
L6: Cs-(Cds-Cd)CtSsH
L7: Cs-(Cds-Cds)CtSsH
L7: Cs-(Cds-Cdd)CtSsH
L8: Cs-(Cds-Cdd-S2d)CtSsH
L8: Cs-(Cds-Cdd-Cd)CtSsH
L5: Cs-CtCtSsH
L5: Cs-CbCsSsH
L5: Cs-CbCdsSsH
L6: Cs-(Cds-Cd)CbSsH
L7: Cs-(Cds-Cds)CbSsH
L7: Cs-(Cds-Cdd)CbSsH
L8: Cs-(Cds-Cdd-S2d)CbSsH
L8: Cs-(Cds-Cdd-Cd)CbSsH
L5: Cs-CbCtSsH
L5: Cs-CbCbSsH
L5: Cs-C=SCbSsH
L5: Cs-C=SC=SSsH
L5: Cs-C=SCsSsH
L5: Cs-C=SCtSsH
L5: Cs-C=S(Cds-Cd)SsH
L6: Cs-C=S(Cds-Cdd)SsH
L7: Cs-C=S(Cds-Cdd-Cd)SsH
L7: Cs-C=S(Cds-Cdd-S2d)SsH
L6: Cs-C=S(Cds-Cds)SsH
L4: Cs-CSsHH
L5: Cs-CsSsHH
L5: Cs-CdsSsHH
L6: Cs-(Cds-Cd)SsHH
L7: Cs-(Cds-Cds)SsHH
L7: Cs-(Cds-Cdd)SsHH
L8: Cs-(Cds-Cdd-S2d)SsHH
L8: Cs-(Cds-Cdd-Cd)SsHH
L5: Cs-CtSsHH
L5: Cs-CbSsHH
L5: Cs-C=SSsHH
L2: O
L3: O2d
L4: O2d-Cd
L4: O2d-O2d
L4: O2d-N3d
L4: O2d-N5dc
L3: O2s
L4: O2s-N
L5: O2s-CN
L6: O2s-CsN3s
L6: O2s-CsN3d
L7: O2s-Cs(N3dOd)
L6: O2s-CdN3d
L7: O2s-(Cd-Cd)(N3dOd)
L6: O2s-CsN5d
L7: O2s-Cs(N5dOdOs)
L6: O2s-CdN5d
L7: O2s-(Cd-CdHH)(N5dOdOs)
L5: O2s-ON
L6: O2s-OsN3s
L6: O2s-OsN3d
L7: O2s-O2s(N3dOd)
L5: O2s-NN
L6: O2s-N3sN3s
L6: O2s-N3sN3d
L7: O2s-N3s(N3dOd)
L4: O2s-HH
L4: O2s-OsH
L4: O2s-OsOs
L4: O2s-CH
L5: O2s-CtH
L5: O2s-CdsH
L6: O2s-(Cds-O2d)H
L6: O2s-(Cds-Cd)H
L5: O2s-CsH
L5: O2s-CbH
L5: O2s-CSH
L4: O2s-OsC
L5: O2s-OsCt
L5: O2s-OsCds
L6: O2s-O2s(Cds-O2d)
L6: O2s-O2s(Cds-Cd)
L5: O2s-OsCs
L5: O2s-OsCb
L4: O2s-CC
L5: O2s-CtCt
L5: O2s-CtCds
L6: O2s-Ct(Cds-O2d)
L6: O2s-Ct(Cds-Cd)
L5: O2s-CtCs
L6: O2s-Cs(CtN3t)
L5: O2s-CtCb
L5: O2s-CdsCds
L6: O2s-(Cds-O2d)(Cds-O2d)
L6: O2s-(Cds-O2d)(Cds-Cd)
L6: O2s-(Cds-Cd)(Cds-Cd)
L5: O2s-CdsCs
L6: O2s-Cs(Cds-O2d)
L6: O2s-Cs(Cds-Cd)
L5: O2s-CdsCb
L6: O2s-Cb(Cds-O2d)
L6: O2s-Cb(Cds-Cd)
L5: O2s-CsCs
L5: O2s-CsCb
L5: O2s-CbCb
L5: O2s-Cs(Cds-S2d)
L2: Si
L2: S
L3: S2d
L4: S2d-Cd
L4: S2d-S2d
L3: S2s
L4: S2s-HH
L4: S2s-CH
L5: S2s-CsH
L5: S2s-CdH
L5: S2s-CtH
L5: S2s-CbH
L5: S2s-COH
L5: S2s-C=SH
L4: S2s-SsH
L4: S2s-SsSs
L4: S2s-SsC
L5: S2s-SsCs
L5: S2s-SsCd
L5: S2s-SsCt
L5: S2s-SsCb
L5: S2s-C=SSs
L4: S2s-CC
L5: S2s-CsCs
L5: S2s-CsCd
L5: S2s-CsCO
L5: S2s-CsCt
L5: S2s-CsCb
L5: S2s-CdCd
L5: S2s-CdCt
L5: S2s-CdCb
L5: S2s-CtCt
L5: S2s-CtCb
L5: S2s-CbCb
L5: S2s-C=SCs
L5: S2s-C=SCt
L5: S2s-C=SC=S
L5: S2s-C=SCd
L5: S2s-C=SCb
L2: N
L3: N1dc
L3: N3s
L4: N3s-CHH
L5: N3s-CsHH
L5: N3s-CbHH
L5: N3s-(CO)HH
L5: N3s-CdHH
L4: N3s-CCH
L5: N3s-CsCsH
L5: N3s-CbCsH
L5: N3s-CbCbH
L5: N3s-(CO)CsH
L5: N3s-(CO)CbH
L5: N3s-(CO)(CO)H
L5: N3s-(CtN3t)CsH
L5: N3s-(CdCd)CsH
L4: N3s-CCC
L5: N3s-CsCsCs
L5: N3s-CbCsCs
L5: N3s-(CO)CsCs
L5: N3s-(CO)(CO)Cs
L5: N3s-(CO)(CO)Cb
L5: N3s-(CtN3t)CsCs
L5: N3s-(CdCd)CsCs
L4: N3s-N3sHH
L4: N3s-NCH
L5: N3s-N3sCsH
L5: N3s-N3sCbH
L5: N3s-CsH(N3dOd)
L5: N3s-CsH(N5dOdOs)
L5: N3s-(CdCd)HN3s
L4: N3s-NCC
L5: N3s-NCsCs
L6: N3s-CsCsN3s
L6: N3s-CsCs(N3dOd)
L6: N3s-CsCs(N5dOdOs)
L5: N3s-NCdCs
L6: N3s-(CdCd)CsN3s
L4: N3s-CsHOs
L4: N3s-CsCsOs
L4: N3s-OsHH
L3: N3d
L4: N3d-CdH
L4: N3d-N3dH
L4: N3d-N3dN3s
L4: N3d-OdOs
L4: N3d-OdN3s
L4: N3d-CsR
L5: N3d-OdC
L5: N3d-CdCs
L5: N3d-N3dCs
L4: N3d-CbR
L3: N5dc
L4: N5dc-OdOsCs
L4: N5dc-OdOsCd
L4: N5dc-OdOsOs
L4: N5dc-OdOsN3s
L3: N5ddc
"""
)
| 18.390346 | 120 | 0.431145 | 161,696 | 853,073 | 2.274626 | 0.013847 | 0.058932 | 0.072116 | 0.033459 | 0.905622 | 0.882612 | 0.857571 | 0.837655 | 0.807389 | 0.79256 | 0 | 0.165521 | 0.319701 | 853,073 | 46,386 | 121 | 18.390743 | 0.468237 | 0.000042 | 0 | 0.611813 | 0 | 0.005942 | 0.21866 | 0.106877 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
d175957c0538886602a21e033a2262d8ddcf487d | 429 | py | Python | Pi to the nth/Pi.py | huffman442/Projects | 16dfcd9bff6e283d6113ed6ebda3f8127576457a | [
"MIT"
] | null | null | null | Pi to the nth/Pi.py | huffman442/Projects | 16dfcd9bff6e283d6113ed6ebda3f8127576457a | [
"MIT"
] | null | null | null | Pi to the nth/Pi.py | huffman442/Projects | 16dfcd9bff6e283d6113ed6ebda3f8127576457a | [
"MIT"
] | null | null | null | import sys
print("How many decimal places would you like to calculate pi to(up to 250 places)?")
digits = input() + 2
pi = "3.1415926535897932384626433832795028841971693993751058209749445923078164062862089986280348253421170679821480865132823066470938446095505822317253594081284811174502841027019385211055596446229489549303819644288109756659334461284756482337867831652712019091"
for x in range(digits):
sys.stdout.write(pi[x]) | 71.5 | 259 | 0.869464 | 34 | 429 | 10.970588 | 0.764706 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.642317 | 0.074592 | 429 | 6 | 260 | 71.5 | 0.297229 | 0 | 0 | 0 | 0 | 0 | 0.762791 | 0.586047 | 0 | 1 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.166667 | 0 | 0.166667 | 0.166667 | 0 | 0 | 1 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | null | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
0f1d36f35254ddeffdef80a6fff0825f7b6ac405 | 352 | py | Python | Pikachurun.py | kartikay22/Pikachu-X | a8bc32192bbfdf0af5dc9ced1194901e4301c7da | [
"MIT"
] | null | null | null | Pikachurun.py | kartikay22/Pikachu-X | a8bc32192bbfdf0af5dc9ced1194901e4301c7da | [
"MIT"
] | null | null | null | Pikachurun.py | kartikay22/Pikachu-X | a8bc32192bbfdf0af5dc9ced1194901e4301c7da | [
"MIT"
] | null | null | null | ##os.system("git clone https://github.com/kartikay22/Pikachu-X /root/userbot && mkdir /root/userbot/bin/ && cd /root/userbot/ && chmod +x /usr/local/bin/* && python3 -m userbot")
import os
os.system("git clone https://github.com/kartikay22/Pikachu-X && mkdir /root/userbot/bin/ && cd /root/userbot/ && chmod +x /usr/local/bin/* && python3 -m userbot")
| 88 | 178 | 0.690341 | 54 | 352 | 4.5 | 0.388889 | 0.226337 | 0.090535 | 0.131687 | 0.921811 | 0.921811 | 0.921811 | 0.921811 | 0.921811 | 0.921811 | 0 | 0.019048 | 0.105114 | 352 | 3 | 179 | 117.333333 | 0.752381 | 0.5 | 0 | 0 | 0 | 0.5 | 0.856322 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 0.5 | 0 | 0.5 | 0 | 0 | 0 | 0 | null | 1 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 13 |
0f57cb2f34e331427865a21c3c25833c3498f6cf | 87 | py | Python | Problems/Working with strings/task.py | gabrielizalo/jetbrains-academy-python-credit-calculator | 093bffde653510348adb35188ecf5cb99e11e73e | [
"MIT"
] | null | null | null | Problems/Working with strings/task.py | gabrielizalo/jetbrains-academy-python-credit-calculator | 093bffde653510348adb35188ecf5cb99e11e73e | [
"MIT"
] | null | null | null | Problems/Working with strings/task.py | gabrielizalo/jetbrains-academy-python-credit-calculator | 093bffde653510348adb35188ecf5cb99e11e73e | [
"MIT"
] | null | null | null | # put your code here
import string
print(string.digits)
print(string.ascii_lowercase)
| 14.5 | 29 | 0.804598 | 13 | 87 | 5.307692 | 0.769231 | 0.318841 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.114943 | 87 | 5 | 30 | 17.4 | 0.896104 | 0.206897 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.2 | 0 | 1 | 0 | true | 0 | 0.333333 | 0 | 0.333333 | 0.666667 | 1 | 0 | 0 | null | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 1 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 0 | 1 | 0 | 7 |
7e472525076f3ec3aae3c5ce6347ec33c5dfd7de | 190,630 | py | Python | backend/tracim_backend/tests/functional/test_contents.py | lezardrouge/tracim | 713ff6066767554333e7e0b1de608ec1a7e4229c | [
"MIT"
] | null | null | null | backend/tracim_backend/tests/functional/test_contents.py | lezardrouge/tracim | 713ff6066767554333e7e0b1de608ec1a7e4229c | [
"MIT"
] | null | null | null | backend/tracim_backend/tests/functional/test_contents.py | lezardrouge/tracim | 713ff6066767554333e7e0b1de608ec1a7e4229c | [
"MIT"
] | null | null | null | # -*- coding: utf-8 -*-
import io
from urllib.parse import quote
from PIL import Image
from depot.io.utils import FileIntent
import pytest
import transaction
from tracim_backend.app_models.contents import content_type_list
from tracim_backend.error import ErrorCode
from tracim_backend.fixtures.content import Content as ContentFixtures
from tracim_backend.fixtures.users_and_groups import Base as BaseFixture
from tracim_backend.lib.core.content import ContentApi
from tracim_backend.lib.core.workspace import WorkspaceApi
from tracim_backend.models.auth import User
from tracim_backend.models.revision_protection import new_revision
from tracim_backend.models.setup_models import get_tm_session
from tracim_backend.tests import FunctionalTest
from tracim_backend.tests import create_1000px_png_test_image
from tracim_backend.tests import set_html_document_slug_to_legacy
class TestFolder(FunctionalTest):
"""
Tests for /api/v2/workspaces/{workspace_id}/folders/{content_id}
endpoint
"""
fixtures = [BaseFixture]
def _setup_basics(self) -> None:
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.dbsession = get_tm_session(self.session_factory, transaction.manager)
self.admin = self.dbsession.query(User).filter(User.email == "admin@admin.admin").one()
self.workspace_api = WorkspaceApi(
current_user=self.admin, session=self.dbsession, config=self.app_config
)
self.content_api = ContentApi(
current_user=self.admin, session=self.dbsession, config=self.app_config
)
self.workspace = self.workspace_api.create_workspace(label="test", save_now=True)
self.folder = self.content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=self.workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
def test_api__get_folder__ok_200__nominal_case(self) -> None:
"""
Get one folder content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test-folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "test-folder"
assert content["parent_id"] is None
assert content["show_in_ui"] is True
assert content["slug"] == "test-folder"
assert content["status"] == "open"
assert content["workspace_id"] == test_workspace.workspace_id
assert content["current_revision_id"] == folder.revision_id
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"]["user_id"] == 1
assert content["last_modifier"]["public_name"] == "Global manager"
assert content["last_modifier"]["avatar_url"] is None
assert content["raw_content"] == ""
def test_api__get_folder__err_400__wrong_content_type(self) -> None:
"""
Get one folder of a content content 7 is not folder
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
thread = content_api.create(
label="thread",
content_type_slug=content_type_list.Thread.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=thread.content_id
),
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_TYPE_NOT_ALLOWED
def test_api__get_folder__err_400__content_does_not_exist(self) -> None:
"""
Get one folder content (content 170 does not exist in db)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
transaction.commit()
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/folders/170".format(
workspace_id=test_workspace.workspace_id
),
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND
def test_api__get_folder__err_400__content_not_in_workspace(self) -> None:
"""
Get one folders of a content (content is in another workspace)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
test_workspace2 = workspace_api.create_workspace(label="test2", save_now=True)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace2.workspace_id, content_id=folder.content_id
),
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND
def test_api__get_folder__err_400__workspace_does_not_exist(self) -> None:
"""
Get one folder content (Workspace 40 does not exist)
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/40/folders/{content_id}".format(content_id=folder.content_id),
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.WORKSPACE_NOT_FOUND
def test_api__get_folder__err_400__workspace_id_is_not_int(self) -> None:
"""
Get one folder content, workspace id is not int
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/coucou/folders/{content_id}".format(content_id=folder.content_id),
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.WORKSPACE_INVALID_ID
def test_api__get_folder__err_400__content_id_is_not_int(self) -> None:
"""
Get one folder content, content_id is not int
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/folders/coucou".format(
workspace_id=test_workspace.workspace_id
),
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_INVALID_ID
def test_api__update_folder__err_400__empty_label(self) -> None:
"""
Update(put) one folder content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {
"label": "",
"raw_content": "<p> Le nouveau contenu </p>",
"sub_content_types": [content_type_list.Folder.slug],
}
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=400,
)
# INFO - G.M - 2018-09-10 - Handled by marshmallow schema
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR
def test_api__update_folder__ok_200__nominal_case(self) -> None:
"""
Update(put) one html document of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {
"label": "My New label",
"raw_content": "<p> Le nouveau contenu </p>",
"sub_content_types": [content_type_list.Folder.slug],
}
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] is None
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == test_workspace.workspace_id
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["sub_content_types"] == [content_type_list.Folder.slug]
def test_api__update_folder__err_400__not_modified(self) -> None:
"""
Update(put) one html document of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {
"label": "My New label",
"raw_content": "<p> Le nouveau contenu </p>",
"sub_content_types": [content_type_list.Folder.slug],
}
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] is None
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == test_workspace.workspace_id
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["sub_content_types"] == [content_type_list.Folder.slug]
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.SAME_VALUE_ERROR
def test_api__update_folder__err_400__allowed_content_changed_only(self) -> None:
"""
Update(put) one folder but change only allowed content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {
"label": "My New label",
"raw_content": "<p> Le nouveau contenu </p>",
"sub_content_types": [content_type_list.Folder.slug],
}
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] is None
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == test_workspace.workspace_id
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["sub_content_types"] == [content_type_list.Folder.slug]
params = {
"label": "My New label",
"raw_content": "<p> Le nouveau contenu </p>",
"sub_content_types": [content_type_list.Folder.slug, content_type_list.Thread.slug],
}
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] is None
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == test_workspace.workspace_id
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert set(content["sub_content_types"]) == set(
[content_type_list.Folder.slug, content_type_list.Thread.slug]
)
def test_api__update_folder__err_400__label_changed_only(self) -> None:
"""
Update(put) one folder but change only allowed content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {
"label": "My New label",
"raw_content": "<p> Le nouveau contenu </p>",
"sub_content_types": [content_type_list.Folder.slug],
}
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] is None
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == test_workspace.workspace_id
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["sub_content_types"] == [content_type_list.Folder.slug]
params = {
"label": "My New label 2",
"raw_content": "<p> Le nouveau contenu </p>",
"sub_content_types": [content_type_list.Folder.slug],
}
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label 2"
assert content["parent_id"] is None
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label-2"
assert content["status"] == "open"
assert content["workspace_id"] == test_workspace.workspace_id
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert set(content["sub_content_types"]) == set([content_type_list.Folder.slug])
def test_api__update_folder__err_400__label_already_used(self) -> None:
"""
Update(put) one html document of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
content_api.create(
label="already_used",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {
"label": "already_used",
"raw_content": "<p> Le nouveau contenu </p>",
"sub_content_types": [content_type_list.Folder.slug],
}
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=400,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER
def test_api__get_folder_revisions__ok_200__nominal_case(self) -> None:
"""
Get one html document of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test-folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=folder):
content_api.update_content(
folder, new_label="test-folder-updated", new_content="Just a test"
)
content_api.save(folder)
with new_revision(session=dbsession, tm=transaction.manager, content=folder):
content_api.archive(folder)
content_api.save(folder)
with new_revision(session=dbsession, tm=transaction.manager, content=folder):
content_api.unarchive(folder)
content_api.save(folder)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}/revisions".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
status=200,
)
revisions = res.json_body
assert len(revisions) == 4
revision = revisions[0]
assert revision["content_type"] == "folder"
assert revision["content_id"] == folder.content_id
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is False
assert revision["label"] == "test-folder"
assert revision["parent_id"] is None
assert revision["show_in_ui"] is True
assert revision["slug"] == "test-folder"
assert revision["status"] == "open"
assert revision["workspace_id"] == test_workspace.workspace_id
assert revision["revision_id"]
assert revision["revision_type"] == "creation"
assert revision["sub_content_types"]
# TODO - G.M - 2018-06-173 - Test with real comments
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 1
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Global manager"
revision = revisions[1]
assert revision["content_type"] == "folder"
assert revision["content_id"] == folder.content_id
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is False
assert revision["label"] == "test-folder-updated"
assert revision["parent_id"] is None
assert revision["show_in_ui"] is True
assert revision["slug"] == "test-folder-updated"
assert revision["status"] == "open"
assert revision["workspace_id"] == test_workspace.workspace_id
assert revision["revision_id"]
assert revision["revision_type"] == "edition"
assert revision["sub_content_types"]
# TODO - G.M - 2018-06-173 - Test with real comments
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 1
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Global manager"
revision = revisions[2]
assert revision["content_type"] == "folder"
assert revision["content_id"] == folder.content_id
assert revision["is_archived"] is True
assert revision["is_deleted"] is False
assert revision["is_editable"] is False
assert revision["label"] != "test-folder-updated"
assert revision["label"].startswith("test-folder-updated")
assert revision["parent_id"] is None
assert revision["show_in_ui"] is True
assert revision["slug"] != "test-folder-updated"
assert revision["slug"].startswith("test-folder-updated")
assert revision["status"] == "open"
assert revision["workspace_id"] == test_workspace.workspace_id
assert revision["revision_id"]
assert revision["revision_type"] == "archiving"
assert revision["sub_content_types"]
# TODO - G.M - 2018-06-173 - Test with real comments
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 1
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Global manager"
revision = revisions[3]
assert revision["content_type"] == "folder"
assert revision["content_id"] == folder.content_id
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is True
assert revision["label"].startswith("test-folder-updated")
assert revision["parent_id"] is None
assert revision["show_in_ui"] is True
assert revision["slug"].startswith("test-folder-updated")
assert revision["status"] == "open"
assert revision["workspace_id"] == test_workspace.workspace_id
assert revision["revision_id"]
assert revision["revision_type"] == "unarchiving"
assert revision["sub_content_types"]
# TODO - G.M - 2018-06-173 - Test with real comments
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 1
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Global manager"
def test_api__set_folder_status__ok_200__nominal_case(self) -> None:
"""
Get one folder content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "closed-deprecated"}
# before
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["status"] == "open"
# set status
self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}/status".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=204,
)
# after
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
status=200,
)
content = res.json_body
assert content["content_type"] == "folder"
assert content["content_id"] == folder.content_id
assert content["status"] == "closed-deprecated"
def test_api__set_folder_status__err_400__wrong_status(self) -> None:
"""
Get one folder content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "unexistant-status"}
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
test_workspace = workspace_api.create_workspace(label="test", save_now=True)
folder = content_api.create(
label="test_folder",
content_type_slug=content_type_list.Folder.slug,
workspace=test_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
res = self.testapp.put_json(
"/api/v2/workspaces/{workspace_id}/folders/{content_id}/status".format(
workspace_id=test_workspace.workspace_id, content_id=folder.content_id
),
params=params,
status=400,
)
# TODO - G.M - 2018-09-10 - handle by marshmallow schema
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR
class TestHtmlDocuments(FunctionalTest):
"""
Tests for /api/v2/workspaces/{workspace_id}/html-documents/{content_id}
endpoint
"""
fixtures = [BaseFixture, ContentFixtures]
def test_api__get_html_document__ok_200__legacy_slug(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
set_html_document_slug_to_legacy(self.session_factory)
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "Tiramisu Recipe"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "tiramisu-recipe"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 27
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] != content["author"]
assert content["last_modifier"]["user_id"] == 3
assert content["last_modifier"]["public_name"] == "Bob i."
assert content["last_modifier"]["avatar_url"] is None
assert (
content["raw_content"] == "<p>To cook a great Tiramisu, you need many ingredients.</p>"
)
assert content["file_extension"] == ".document.html"
def test_api__get_html_document__ok_200__nominal_case(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "Tiramisu Recipe"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "tiramisu-recipe"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 27
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] != content["author"]
assert content["last_modifier"]["user_id"] == 3
assert content["last_modifier"]["public_name"] == "Bob i."
assert content["last_modifier"]["avatar_url"] is None
assert (
content["raw_content"] == "<p>To cook a great Tiramisu, you need many ingredients.</p>"
)
assert content["file_extension"] == ".document.html"
def test_api__get_html_document__ok_200__archived_content(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put_json("/api/v2/workspaces/2/contents/6/archived", status=204)
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["is_archived"] is True
def test_api__get_html_document__ok_200__deleted_content(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put_json("/api/v2/workspaces/2/contents/6/trashed", status=204)
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["is_deleted"] is True
def test_api__get_html_document__err_400__wrong_content_type(self) -> None:
"""
Get one html document of a content content 7 is not html_document
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/html-documents/7", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_TYPE_NOT_ALLOWED
def test_api__get_html_document__err_400__content_does_not_exist(self) -> None:
"""
Get one html document of a content (content 170 does not exist in db
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/html-documents/170", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND
def test_api__get_html_document__err_400__content_not_in_workspace(self) -> None:
"""
Get one html document of a content (content 6 is in workspace 2)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/1/html-documents/6", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND
def test_api__get_html_document__err_400__workspace_does_not_exist(self) -> None:
"""
Get one html document of a content (Workspace 40 does not exist)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/40/html-documents/6", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.WORKSPACE_NOT_FOUND
def test_api__get_html_document__err_400__workspace_id_is_not_int(self) -> None:
"""
Get one html document of a content, workspace id is not int
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/coucou/html-documents/6", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.WORKSPACE_INVALID_ID
def test_api__get_html_document__err_400__content_id_is_not_int(self) -> None:
"""
Get one html document of a content, content_id is not int
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/html-documents/coucou", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_INVALID_ID
def test_api__update_html_document__err_400__empty_label(self) -> None:
"""
Update(put) one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6", params=params, status=400
)
# INFO - G.M - 2018-09-10 - Handled by marshmallow schema
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR
def test_api__update_html_document__ok_200__nominal_case(self) -> None:
"""
Update(put) one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6", params=params, status=200
)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 28
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["file_extension"] == ".document.html"
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 28
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["file_extension"] == ".document.html"
def test_api__update_html_document__err_400__not_editable(self) -> None:
"""
Update(put) one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "closed-deprecated"}
self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6/status", params=params, status=204
)
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu ! </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6", params=params, status=400
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE
def test_api__update_html_document__err_400__not_modified(self) -> None:
"""
Update(put) one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6", params=params, status=200
)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 28
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 28
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
res = self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6", params=params, status=400
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.SAME_VALUE_ERROR
def test_api__get_html_document_revisions__ok_200__nominal_case(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6/revisions", status=200)
revisions = res.json_body
assert len(revisions) == 3
revision = revisions[0]
assert revision["content_type"] == "html-document"
assert revision["content_id"] == 6
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is False
assert revision["label"] == "Tiramisu Recipes!!!"
assert revision["parent_id"] == 3
assert revision["show_in_ui"] is True
assert revision["slug"] == "tiramisu-recipes"
assert revision["status"] == "open"
assert revision["workspace_id"] == 2
assert revision["revision_id"] == 6
assert revision["revision_type"] == "creation"
assert revision["sub_content_types"]
# TODO - G.M - 2018-06-173 - Test with real comments
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 1
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Global manager"
revision = revisions[1]
assert revision["content_type"] == "html-document"
assert revision["content_id"] == 6
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is False
assert revision["label"] == "Tiramisu Recipes!!!"
assert revision["parent_id"] == 3
assert revision["show_in_ui"] is True
assert revision["slug"] == "tiramisu-recipes"
assert revision["status"] == "open"
assert revision["workspace_id"] == 2
assert revision["revision_id"] == 7
assert revision["revision_type"] == "edition"
assert revision["sub_content_types"]
# TODO - G.M - 2018-06-173 - Test with real comments
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 1
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Global manager"
revision = revisions[2]
assert revision["content_type"] == "html-document"
assert revision["content_id"] == 6
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is True
assert revision["label"] == "Tiramisu Recipe"
assert revision["parent_id"] == 3
assert revision["show_in_ui"] is True
assert revision["slug"] == "tiramisu-recipe"
assert revision["status"] == "open"
assert revision["workspace_id"] == 2
assert revision["revision_id"] == 27
assert revision["revision_type"] == "edition"
assert revision["sub_content_types"]
# TODO - G.M - 2018-06-173 - Test with real comments
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 3
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Bob i."
assert revision["file_extension"] == ".document.html"
def test_api__set_html_document_status__ok_200__nominal_case(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "closed-deprecated"}
# before
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["status"] == "open"
# set status
self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6/status", params=params, status=204
)
# after
res = self.testapp.get("/api/v2/workspaces/2/html-documents/6", status=200)
content = res.json_body
assert content["content_type"] == "html-document"
assert content["content_id"] == 6
assert content["status"] == "closed-deprecated"
def test_api__set_html_document_status__err_400__wrong_status(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "unexistant-status"}
res = self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6/status", params=params, status=400
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR
def test_api__set_document_status__err_400__same_status(self) -> None:
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "open"}
res = self.testapp.put_json(
"/api/v2/workspaces/2/html-documents/6/status", params=params, status=400
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.INVALID_STATUS_CHANGE
class TestFiles(FunctionalTest):
"""
Tests for /api/v2/workspaces/{workspace_id}/files/{content_id}
endpoint
"""
fixtures = [BaseFixture, ContentFixtures]
def test_api__get_file__ok_200__nominal_case(self) -> None:
"""
Get one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file"
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "Test_file"
assert content["parent_id"] == 1
assert content["show_in_ui"] is True
assert content["slug"] == "test-file"
assert content["status"] == "open"
assert content["workspace_id"] == 1
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p>description</p>"
assert content["mimetype"] == "plain/text"
assert content["size"] == len(b"Test file")
assert content["file_extension"] == ".txt"
assert content["filename"] == "Test_file.txt"
assert content["page_nb"] == 1
assert content["has_pdf_preview"] is True
assert content["has_jpeg_preview"] is True
def test_api__get_file__ok_200__no_file_add(self) -> None:
"""
Get one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "Test file"
assert content["parent_id"] == 1
assert content["show_in_ui"] is True
assert content["slug"] == "test-file"
assert content["status"] == "open"
assert content["workspace_id"] == 1
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == ""
assert content["mimetype"] == ""
assert content["file_extension"] == ""
assert content["filename"] == "Test file"
assert content["size"] is None
assert content["page_nb"] is None
assert content["has_pdf_preview"] is False
assert content["has_jpeg_preview"] is False
def test_api__get_file__ok_200__binary_file(self) -> None:
"""
Get one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file,
"Test_file.bin",
new_mimetype="application/octet-stream",
new_content=bytes(100),
)
content_api.save(test_file)
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "Test_file"
assert content["parent_id"] == 1
assert content["show_in_ui"] is True
assert content["slug"] == "test-file"
assert content["status"] == "open"
assert content["workspace_id"] == 1
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == ""
assert content["mimetype"] == "application/octet-stream"
assert content["size"] == 100
assert content["file_extension"] == ".bin"
assert content["filename"] == "Test_file.bin"
assert content["page_nb"] is None
assert content["has_pdf_preview"] is False
assert content["has_jpeg_preview"] is False
def test_api__get_files__err_400__wrong_content_type(self) -> None:
"""
Get one file of a content content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/files/6", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_TYPE_NOT_ALLOWED
def test_api__get_file__err_400__content_does_not_exist(self) -> None:
"""
Get one file (content 170 does not exist in db
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/1/files/170", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND
def test_api__get_file__err_400__content_not_in_workspace(self) -> None:
"""
Get one file (content 9 is in workspace 2)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/1/files/9", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND
def test_api__get_file__err_400__workspace_does_not_exist(self) -> None:
"""
Get one file (Workspace 40 does not exist)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/40/files/9", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.WORKSPACE_NOT_FOUND
def test_api__get_file__err_400__workspace_id_is_not_int(self) -> None:
"""
Get one file, workspace id is not int
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/coucou/files/9", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.WORKSPACE_INVALID_ID
def test_api__get_file__err_400__content_id_is_not_int(self) -> None:
"""
Get one file, content_id is not int
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/files/coucou", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_INVALID_ID
def test_api__update_file_info_err_400__empty_label(self) -> None:
"""
Update(put) one file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file"
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400
)
# INFO - G.M - 2018-09-10 - Handle by marshmallow schema
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR
def test_api__update_file_info__ok_200__nominal_case(self) -> None:
"""
Update(put) one file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
with dbsession.no_autoflush:
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
content_api.update_file_data(
test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file"
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 1
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 1
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["mimetype"] == "plain/text"
assert content["size"] == len(b"Test file")
assert content["page_nb"] == 1
assert content["has_pdf_preview"] is True
assert content["has_jpeg_preview"] is True
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 1
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 1
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["mimetype"] == "plain/text"
assert content["size"] == len(b"Test file")
assert content["page_nb"] == 1
assert content["has_pdf_preview"] is True
assert content["has_jpeg_preview"] is True
def test_api__update_file_info__err_400__content_status_closed(self) -> None:
"""
Update(put) one file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
with dbsession.no_autoflush:
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
content_api.update_file_data(
test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file"
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
test_file.status = "closed-validated"
content_api.save(test_file)
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE
def test_api__update_file_info__err_400__content_deleted(self) -> None:
"""
Update(put) one file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(
current_user=admin, session=dbsession, config=self.app_config, show_deleted=True
)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
with dbsession.no_autoflush:
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
content_api.update_file_data(
test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file"
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
test_file.is_deleted = True
content_api.save(test_file)
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE
def test_api__update_file_info__err_400__content_archived(self) -> None:
"""
Update(put) one file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(
current_user=admin, session=dbsession, config=self.app_config, show_deleted=True
)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
with dbsession.no_autoflush:
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
content_api.update_file_data(
test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file"
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
test_file.is_archived = True
content_api.save(test_file)
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE
def test_api__update_file_info__err_400__not_modified(self) -> None:
"""
Update(put) one file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file"
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 1
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 1
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["mimetype"] == "plain/text"
assert content["size"] == len(b"Test file")
assert content["page_nb"] == 1
assert content["has_pdf_preview"] is True
assert content["has_jpeg_preview"] is True
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 1
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 1
assert content["current_revision_id"]
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["mimetype"] == "plain/text"
assert content["size"] == len(b"Test file")
assert content["page_nb"] == 1
assert content["has_pdf_preview"] is True
assert content["has_jpeg_preview"] is True
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.SAME_VALUE_ERROR
def test_api__update_file_info__err_400__label_already_used(self) -> None:
"""
Update(put) one file, failed because label already used
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="folder_used",
do_save=True,
do_notify=False,
)
with dbsession.no_autoflush:
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
test_file2 = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
filename="already_used.txt",
do_save=False,
do_notify=False,
)
test_file2.file_extension = ".txt"
test_file2.depot_file = FileIntent(b"Test file", "already_used.txt", "text/plain")
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "folder_used", "raw_content": "<p> Le nouveau contenu </p>"}
self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=200
)
params = {"label": "already_used", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), params=params, status=400
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER
def test_api__get_file_revisions__ok_200__nominal_case(self) -> None:
"""
Get file revisions
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file, "Test_file.txt", new_mimetype="plain/text", new_content=b"Test file"
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/revisions".format(test_file.content_id), status=200
)
revisions = res.json_body
assert len(revisions) == 1
revision = revisions[0]
assert revision["content_type"] == "file"
assert revision["content_id"] == test_file.content_id
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is True
assert revision["label"] == "Test_file"
assert revision["parent_id"] == 1
assert revision["show_in_ui"] is True
assert revision["slug"] == "test-file"
assert revision["status"] == "open"
assert revision["workspace_id"] == 1
assert revision["revision_id"]
assert revision["sub_content_types"]
# TODO - G.M - 2018-06-173 - Test with real comments
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 1
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Global manager"
assert revision["mimetype"] == "plain/text"
assert revision["size"] == len(b"Test file")
assert revision["page_nb"] == 1
assert revision["has_pdf_preview"] is True
def test_api__set_file_status__ok_200__nominal_case(self) -> None:
"""
set file status
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "closed-deprecated"}
# before
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["status"] == "open"
# set status
self.testapp.put_json(
"/api/v2/workspaces/1/files/{}/status".format(test_file.content_id),
params=params,
status=204,
)
# after
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["status"] == "closed-deprecated"
def test_api__set_file_status__err_400__wrong_status(self) -> None:
"""
set file status
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "unexistant-status"}
# before
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}".format(test_file.content_id), status=200
)
content = res.json_body
assert content["content_type"] == "file"
assert content["content_id"] == test_file.content_id
assert content["status"] == "open"
# set status
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}/status".format(test_file.content_id),
params=params,
status=400,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR
def test_api__get_file_raw__ok_200__nominal_case(self) -> None:
"""
Get one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
filename = "Test_file.txt"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, filename), status=200
)
assert res.body == b"Test file"
assert res.content_type == "text/plain"
assert res.content_length == len(b"Test file")
assert int(res.headers["Content-Length"]) == res.content_length
assert res.last_modified.second == test_file.updated.second
assert res.last_modified.minute == test_file.updated.minute
assert res.last_modified.day == test_file.updated.day
assert res.last_modified.month == test_file.updated.month
assert res.last_modified.year == test_file.updated.year
def test_api__get_file_raw__ok_200__force_download_case(self) -> None:
"""
Get one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file,
new_content=b"Test file",
new_filename="Test_file.txt",
new_mimetype="text/plain",
)
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"force_download": 1}
filename = "Test_file.txt"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, filename),
status=200,
params=params,
)
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename)
assert res.body == b"Test file"
assert res.content_type == "text/plain"
assert res.content_length == len(b"Test file")
assert int(res.headers["Content-Length"]) == res.content_length
assert res.last_modified.second == test_file.updated.second
assert res.last_modified.minute == test_file.updated.minute
assert res.last_modified.day == test_file.updated.day
assert res.last_modified.month == test_file.updated.month
assert res.last_modified.year == test_file.updated.year
def test_api__create_file__ok__200__nominal_case(self) -> None:
"""
create one file of a content at workspace root
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
image = create_1000px_png_test_image()
res = self.testapp.post(
"/api/v2/workspaces/{}/files".format(business_workspace.workspace_id),
upload_files=[("files", image.name, image.getvalue())],
status=200,
)
res = res.json_body
assert res["parent_id"] is None
assert res["content_type"] == "file"
assert res["is_archived"] is False
assert res["is_deleted"] is False
assert res["is_editable"] is True
assert res["workspace_id"] == business_workspace.workspace_id
assert isinstance(res["content_id"], int)
content_id = res["content_id"]
assert res["status"] == "open"
assert res["label"] == "test_image"
assert res["slug"] == "test-image"
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/files/{content_id}".format(
workspace_id=business_workspace.workspace_id, content_id=content_id
),
status=200,
)
res = res.json_body
assert res["parent_id"] is None
assert res["content_type"] == "file"
assert res["is_archived"] is False
assert res["is_deleted"] is False
assert res["is_editable"] is True
assert res["workspace_id"] == business_workspace.workspace_id
assert isinstance(res["content_id"], int)
assert res["status"] == "open"
assert res["label"] == "test_image"
assert res["slug"] == "test-image"
assert res["author"]["user_id"] == admin.user_id
assert res["page_nb"] == 1
assert res["mimetype"] == "image/png"
def test_api__create_file__err_400__filename_already_used(self) -> None:
"""
create one file of a content but filename is already used here
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
image = create_1000px_png_test_image()
res = self.testapp.post(
"/api/v2/workspaces/{}/files".format(business_workspace.workspace_id),
upload_files=[("files", image.name, image.getvalue())],
status=200,
)
res = res.json_body
assert res["parent_id"] is None
assert res["content_type"] == "file"
assert res["is_archived"] is False
assert res["is_deleted"] is False
assert res["is_editable"] is True
assert res["workspace_id"] == business_workspace.workspace_id
assert isinstance(res["content_id"], int)
assert res["status"] == "open"
assert res["label"] == "test_image"
assert res["slug"] == "test-image"
res = self.testapp.post(
"/api/v2/workspaces/{}/files".format(business_workspace.workspace_id),
upload_files=[("files", image.name, image.getvalue())],
status=400,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER
def test_api__create_file__ok__200__in_folder(self) -> None:
"""
create one file of a content in a folder
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
folder = content_api.create(
label="test-folder",
content_type_slug=content_type_list.Folder.slug,
workspace=business_workspace,
do_save=True,
do_notify=False,
)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"parent_id": folder.content_id}
image = create_1000px_png_test_image()
res = self.testapp.post(
"/api/v2/workspaces/{}/files".format(business_workspace.workspace_id),
upload_files=[("files", image.name, image.getvalue())],
params=params,
status=200,
)
res = res.json_body
assert res["parent_id"] == folder.content_id
assert res["content_type"] == "file"
assert res["is_archived"] is False
assert res["is_deleted"] is False
assert res["is_editable"] is True
assert res["workspace_id"] == business_workspace.workspace_id
assert isinstance(res["content_id"], int)
content_id = res["content_id"]
assert res["status"] == "open"
assert res["label"] == "test_image"
assert res["slug"] == "test-image"
res = self.testapp.get(
"/api/v2/workspaces/{workspace_id}/files/{content_id}".format(
workspace_id=business_workspace.workspace_id, content_id=content_id
),
status=200,
)
res = res.json_body
assert res["parent_id"] == folder.content_id
assert res["content_type"] == "file"
assert res["is_archived"] is False
assert res["is_deleted"] is False
assert res["is_editable"] is True
assert res["workspace_id"] == business_workspace.workspace_id
assert isinstance(res["content_id"], int)
assert res["status"] == "open"
assert res["label"] == "test_image"
assert res["slug"] == "test-image"
assert res["author"]["user_id"] == admin.user_id
assert res["page_nb"] == 1
assert res["mimetype"] == "image/png"
def test_api__create_file__err__400__unallow_subcontent(self) -> None:
"""
create one file of a content but subcontent of type file unallowed here
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
folder = content_api.create(
label="test-folder",
content_type_slug=content_type_list.Folder.slug,
workspace=business_workspace,
do_save=True,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=folder):
content_api.set_allowed_content(folder, [])
content_api.save(folder)
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"parent_id": folder.content_id}
image = create_1000px_png_test_image()
res = self.testapp.post(
"/api/v2/workspaces/{}/files".format(business_workspace.workspace_id),
upload_files=[("files", image.name, image.getvalue())],
params=params,
status=400,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.UNALLOWED_SUBCONTENT
def test_api__create_file__err__400__parent_not_found(self) -> None:
"""
create one file of a content but parent_id is not valid
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"parent_id": 3000}
image = create_1000px_png_test_image()
res = self.testapp.post(
"/api/v2/workspaces/{}/files".format(business_workspace.workspace_id),
upload_files=[("files", image.name, image.getvalue())],
params=params,
status=400,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.PARENT_NOT_FOUND
def test_api__set_file_raw__ok_200__nominal_case(self) -> None:
"""
Set one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), status=200
)
assert res.body == image.getvalue()
assert res.content_type == "image/png"
assert res.content_length == len(image.getvalue())
def test_api__set_file_raw__ok_200__filename_already_used(self) -> None:
"""
Set one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file_2 = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file2",
do_save=False,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
content2_id = int(test_file_2.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content2_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=400,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER
def test_api__set_file_raw__err_400__closed_status_file(self) -> None:
"""
Set one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.status = "closed-validated"
content_api.save(test_file)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=400,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_IN_NOT_EDITABLE_STATE
@pytest.mark.xfail(raises=AssertionError, reason="Broken feature dues to pyramid behaviour")
def test_api__set_file_raw__err_400_not_modified(self) -> None:
"""
Set one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name), status=200
)
assert res.body == image.getvalue()
assert res.content_type == "image/png"
assert res.content_length == len(image.getvalue())
res = self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status="*",
)
assert res.status == 400
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.CONTENT_FILENAME_ALREADY_USED_IN_FOLDER
def test_api__get_allowed_size_dim__ok__nominal_case(self) -> None:
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
content_id = int(test_file.content_id)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/allowed_dims".format(content_id), status=200
)
res = res.json_body
assert res["restricted"] is True
assert len(res["dimensions"]) == 1
dim = res["dimensions"][0]
assert dim["width"] == 256
assert dim["height"] == 256
def test_api__get_jpeg_preview__ok__200__nominal_case(self) -> None:
"""
Set one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/".format(content_id), status=200
)
assert res.body != image.getvalue()
assert res.content_type == "image/jpeg"
def test_api__get_jpeg_preview__ok__200__force_download_case(self) -> None:
"""
Set one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
params = {"force_download": 1}
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/raw".format(content_id),
status=200,
params=params,
)
filename = "test_image_page_1.jpg"
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename)
assert res.body != image.getvalue()
assert res.content_type == "image/jpeg"
def test_api__get_jpeg_preview__err_400__UnavailablePreview(self) -> None:
"""
Set one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file,
"Test_file.bin",
new_mimetype="application/octet-stream",
new_content=bytes(100),
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"force_download": 0}
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/".format(content_id),
status=400,
params=params,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.UNAIVALABLE_PREVIEW
def test_api__get_sized_jpeg_preview__ok__200__nominal_case(self) -> None:
"""
get 256x256 preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/256x256/{}".format(content_id, image.name),
status=200,
)
assert res.body != image.getvalue()
assert res.content_type == "image/jpeg"
new_image = Image.open(io.BytesIO(res.body))
assert 256, 256 == new_image.size
def test_api__get_sized_jpeg_preview__err_400__UnavailablePreview(self) -> None:
"""
Set one file of a content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file,
"Test_file.bin",
new_mimetype="application/octet-stream",
new_content=bytes(100),
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"force_download": 0}
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/256x256/{}".format(
content_id, "Test_file.bin"
),
status=400,
params=params,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.UNAIVALABLE_PREVIEW
def test_api__get_sized_jpeg_preview__ok__200__force_download_case(self) -> None:
"""
get 256x256 preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
params = {"force_download": 1}
dl_filename = "test_image_page_1_256x256.jpg"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/256x256/{}".format(content_id, dl_filename),
status=200,
params=params,
)
assert res.body != image.getvalue()
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(dl_filename, dl_filename)
assert res.content_type == "image/jpeg"
new_image = Image.open(io.BytesIO(res.body))
assert 256, 256 == new_image.size
def test_api__get_sized_jpeg_preview__ok__200__force_download_case_no_filename(self) -> None:
"""
get 256x256 preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
params = {"force_download": 1}
dl_filename = "test_image_page_1_256x256.jpg"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/256x256/".format(content_id),
status=200,
params=params,
)
assert res.body != image.getvalue()
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(dl_filename, dl_filename)
assert res.content_type == "image/jpeg"
new_image = Image.open(io.BytesIO(res.body))
assert 256, 256 == new_image.size
def test_api__get_sized_jpeg_preview__ok__200__force_download_case_filename_is_raw(
self
) -> None:
"""
get 256x256 preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
params = {"force_download": 1}
dl_filename = "test_image_page_1_256x256.jpg"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/256x256/raw".format(content_id),
status=200,
params=params,
)
assert res.body != image.getvalue()
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(dl_filename, dl_filename)
assert res.content_type == "image/jpeg"
new_image = Image.open(io.BytesIO(res.body))
assert 256, 256 == new_image.size
def test_api__get_sized_jpeg_preview__err__400__SizeNotAllowed(self) -> None:
"""
get 256x256 preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
filename = "test_image_512x512.jpg"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/jpg/512x512/{}".format(content_id, filename),
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.PREVIEW_DIM_NOT_ALLOWED
def test_api__get_sized_jpeg_revision_preview__ok__200__nominal_case(self) -> None:
"""
get 256x256 revision preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
revision_id = int(test_file.revision_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
filename = "test_file.txt"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format(
content_id=content_id, revision_id=revision_id, filename=filename
),
status=200,
)
assert res.content_type == "text/plain"
filename = "test_image_256x256.jpg"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/jpg/256x256/{filename}".format(
content_id=content_id, revision_id=revision_id, filename=filename
),
status=200,
)
assert res.body != image.getvalue()
assert res.content_type == "image/jpeg"
new_image = Image.open(io.BytesIO(res.body))
assert 256, 256 == new_image.size
def test_api__get_sized_jpeg_revision_preview__ok__200__force_download_case(self) -> None:
"""
get 256x256 revision preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
revision_id = int(test_file.revision_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format(
content_id=content_id, revision_id=revision_id, filename=image.name
),
status=200,
)
assert res.content_type == "text/plain"
params = {"force_download": 1}
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/jpg/256x256/".format(
content_id=content_id, revision_id=revision_id
),
status=200,
params=params,
)
filename = "Test file_r{}_page_1_256x256.jpg".format(revision_id)
urlencoded_filename = quote(filename)
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(
filename, urlencoded_filename
)
assert res.body != image.getvalue()
assert res.content_type == "image/jpeg"
new_image = Image.open(io.BytesIO(res.body))
assert 256, 256 == new_image.size
def test_api__get_full_pdf_preview__ok__200__nominal_case(self) -> None:
"""
get full pdf preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, test_file.file_name),
upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())],
status=204,
)
filename = "test_image.pdf"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, filename),
status=200,
)
assert res.content_type == "application/pdf"
def test_api__get_full_pdf_preview__ok__200__force_download_case(self) -> None:
"""
get full pdf preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
filename = "Test_file.txt"
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, filename),
upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())],
status=204,
)
params = {"force_download": 1}
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, filename),
status=200,
params=params,
)
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename)
assert res.content_type == "application/pdf"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, "Test_file.pdf"),
status=200,
params=params,
)
filename = "Test_file.pdf"
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename)
assert res.content_type == "application/pdf"
def test_api__get_full_pdf_preview__err__400__png_UnavailablePreviewType(self) -> None:
"""
get full pdf preview of a png image -> error UnavailablePreviewType
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, image.name),
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.UNAVAILABLE_PREVIEW_TYPE
def test_api__get_full_pdf_preview__err__400__png_UnavailablePreview(self) -> None:
"""
get full pdf preview of a png image -> error UnavailablePreviewType
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file,
"Test_file.bin",
new_mimetype="application/octet-stream",
new_content=bytes(100),
)
dbsession.flush()
content_id = test_file.content_id
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
filename = "Test_file.bin"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/full/{}".format(content_id, filename),
status=400,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.UNAIVALABLE_PREVIEW
def test_api__get_pdf_preview__ok__200__nominal_case(self) -> None:
"""
get full pdf preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, test_file.file_name),
upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())],
status=204,
)
params = {"page": 1}
filename = "test_file.pdf"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/{}".format(content_id, filename),
status=200,
params=params,
)
assert res.content_type == "application/pdf"
def test_api__get_pdf_preview_err__400__UnavailablePreview(self) -> None:
"""
get full pdf preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_file_data(
test_file,
"Test_file.bin",
new_mimetype="application/octet-stream",
new_content=bytes(100),
)
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"page": 1}
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/".format(content_id),
status=400,
params=params,
)
assert isinstance(res.json, dict)
assert "code" in res.json.keys()
assert res.json_body["code"] == ErrorCode.UNAIVALABLE_PREVIEW
def test_api__get_pdf_preview__ok__200__force_download_case(self) -> None:
"""
get full pdf preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
filename = "test_file.txt"
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, filename),
upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())],
status=204,
)
filename = "Test_file_page_1.pdf"
params = {"page": 1, "force_download": 1}
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/{}".format(content_id, filename),
status=200,
params=params,
)
assert res.content_type == "application/pdf"
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename)
def test_api__get_pdf_preview__ok__err__400_page_of_preview_not_found(self) -> None:
"""
get full pdf preview of a txt file
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=True,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/".format(content_id),
upload_files=[("files", test_file.file_name, test_file.depot_file.file.read())],
status=204,
)
params = {"page": 2}
res = self.testapp.get(
"/api/v2/workspaces/1/files/{}/preview/pdf/".format(content_id),
status=400,
params=params,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.PAGE_OF_PREVIEW_NOT_FOUND
def test_api__get_pdf_revision_preview__ok__200__nominal_case(self) -> None:
"""
get pdf revision preview of content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
revision_id = int(test_file.revision_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
filename = image.name
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format(
content_id=content_id, revision_id=revision_id, filename=filename
),
status=200,
)
assert res.content_type == "text/plain"
params = {"page": 1}
filename = "test_image__page_1.pdf"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/pdf/{filename}".format(
content_id=content_id, revision_id=revision_id, params=params, filename=filename
),
status=200,
)
assert res.content_type == "application/pdf"
def test_api__get_full_pdf_revision_preview__ok__200__nominal_case(self) -> None:
"""
get pdf revision preview of content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
revision_id = int(test_file.revision_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/".format(
content_id=content_id, revision_id=revision_id
),
status=200,
)
assert res.content_type == "text/plain"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/pdf/full/".format(
content_id=content_id, revision_id=revision_id
),
status=200,
)
assert res.content_type == "application/pdf"
def test_api__get_full_pdf_revision_preview__ok__200__force_download_case(self) -> None:
"""
get pdf revision preview of content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
revision_id = int(test_file.revision_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format(
content_id=content_id, revision_id=revision_id, filename=image.name
),
status=200,
)
assert res.content_type == "text/plain"
params = {"force_download": 1}
filename = "Test_file.pdf"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/pdf/full/{filename}".format(
content_id=content_id, revision_id=revision_id, filename="Test_file.pdf"
),
status=200,
params=params,
)
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename)
assert res.content_type == "application/pdf"
def test_api__get_pdf_revision_preview__ok__200__force_download_case(self) -> None:
"""
get pdf revision preview of content
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
dbsession.flush()
transaction.commit()
content_id = int(test_file.content_id)
revision_id = int(test_file.revision_id)
image = create_1000px_png_test_image()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
self.testapp.put(
"/api/v2/workspaces/1/files/{}/raw/{}".format(content_id, image.name),
upload_files=[("files", image.name, image.getvalue())],
status=204,
)
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/raw/{filename}".format(
content_id=content_id, revision_id=revision_id, filename=image.name
),
status=200,
)
assert res.content_type == "text/plain"
params = {"page": 1, "force_download": 1}
filename = "test_image_page_1.pdf"
res = self.testapp.get(
"/api/v2/workspaces/1/files/{content_id}/revisions/{revision_id}/preview/pdf/{filename}".format(
content_id=content_id, revision_id=revision_id, filename=filename
),
status=200,
params=params,
)
assert res.headers[
"Content-Disposition"
] == "attachment; filename=\"{}\"; filename*=UTF-8''{};".format(filename, filename)
assert res.content_type == "application/pdf"
def test_api__set_file_status__err_400__same_status(self) -> None:
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_file = content_api.create(
content_type_slug=content_type_list.File.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test file",
do_save=False,
do_notify=False,
)
test_file.file_extension = ".txt"
test_file.depot_file = FileIntent(b"Test file", "Test_file.txt", "text/plain")
with new_revision(session=dbsession, tm=transaction.manager, content=test_file):
content_api.update_content(test_file, "Test_file", "<p>description</p>")
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "open"}
# set status
res = self.testapp.put_json(
"/api/v2/workspaces/1/files/{}/status".format(test_file.content_id),
params=params,
status=400,
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.INVALID_STATUS_CHANGE
class TestThreads(FunctionalTest):
"""
Tests for /api/v2/workspaces/{workspace_id}/threads/{content_id}
endpoint
"""
fixtures = [BaseFixture, ContentFixtures]
def test_api__get_thread__err_400__wrong_content_type(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/threads/6", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_TYPE_NOT_ALLOWED
def test_api__get_thread__ok_200__nominal_case(self) -> None:
"""
Get one html document of a content
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200)
content = res.json_body
assert content["content_type"] == "thread"
assert content["content_id"] == 7
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "Best Cakes?"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "best-cakes"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 26
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] != content["author"]
assert content["last_modifier"]["user_id"] == 3
assert content["last_modifier"]["public_name"] == "Bob i."
assert content["last_modifier"]["avatar_url"] is None
assert content["raw_content"] == "What is the best cake?"
assert content["file_extension"] == ".thread.html"
assert content["filename"] == "Best Cakes?.thread.html"
def test_api__get_thread__err_400__content_does_not_exist(self) -> None:
"""
Get one thread (content 170 does not exist)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/threads/170", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND
def test_api__get_thread__err_400__content_not_in_workspace(self) -> None:
"""
Get one thread(content 7 is in workspace 2)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/1/threads/7", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_NOT_FOUND
def test_api__get_thread__err_400__workspace_does_not_exist(self) -> None:
"""
Get one thread (Workspace 40 does not exist)
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/40/threads/7", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.WORKSPACE_NOT_FOUND
def test_api__get_thread__err_400__workspace_id_is_not_int(self) -> None:
"""
Get one thread, workspace id is not int
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/coucou/threads/7", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.WORKSPACE_INVALID_ID
def test_api__get_thread__err_400_content_id_is_not_int(self) -> None:
"""
Get one thread, content id is not int
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/threads/coucou", status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.CONTENT_INVALID_ID
def test_api__update_thread__ok_200__nominal_case(self) -> None:
"""
Update(put) thread
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json("/api/v2/workspaces/2/threads/7", params=params, status=200)
content = res.json_body
assert content["content_type"] == "thread"
assert content["content_id"] == 7
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 28
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["file_extension"] == ".thread.html"
assert content["filename"] == "My New label.thread.html"
res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200)
content = res.json_body
assert content["content_type"] == "thread"
assert content["content_id"] == 7
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 28
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
assert content["file_extension"] == ".thread.html"
assert content["filename"] == "My New label.thread.html"
def test_api__update_thread__err_400__not_modified(self) -> None:
"""
Update(put) thread
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "My New label", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json("/api/v2/workspaces/2/threads/7", params=params, status=200)
content = res.json_body
assert content["content_type"] == "thread"
assert content["content_id"] == 7
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 28
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200)
content = res.json_body
assert content["content_type"] == "thread"
assert content["content_id"] == 7
assert content["is_archived"] is False
assert content["is_deleted"] is False
assert content["is_editable"] is True
assert content["label"] == "My New label"
assert content["parent_id"] == 3
assert content["show_in_ui"] is True
assert content["slug"] == "my-new-label"
assert content["status"] == "open"
assert content["workspace_id"] == 2
assert content["current_revision_id"] == 28
# TODO - G.M - 2018-06-173 - check date format
assert content["created"]
assert content["author"]
assert content["author"]["user_id"] == 1
assert content["author"]["avatar_url"] is None
assert content["author"]["public_name"] == "Global manager"
# TODO - G.M - 2018-06-173 - check date format
assert content["modified"]
assert content["last_modifier"] == content["author"]
assert content["raw_content"] == "<p> Le nouveau contenu </p>"
res = self.testapp.put_json("/api/v2/workspaces/2/threads/7", params=params, status=400)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.SAME_VALUE_ERROR
def test_api__update_thread__err_400__empty_label(self) -> None:
"""
Update(put) thread
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"label": "", "raw_content": "<p> Le nouveau contenu </p>"}
res = self.testapp.put_json("/api/v2/workspaces/2/threads/7", params=params, status=400)
# TODO - G.M - 2018-09-10 - Handle by marshmallow schema
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR
def test_api__get_thread_revisions__ok_200__nominal_case(self) -> None:
"""
Get threads revisions
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get("/api/v2/workspaces/2/threads/7/revisions", status=200)
revisions = res.json_body
assert len(revisions) == 2
revision = revisions[0]
assert revision["content_type"] == "thread"
assert revision["content_id"] == 7
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is False
assert revision["label"] == "Best Cake"
assert revision["parent_id"] == 3
assert revision["show_in_ui"] is True
assert revision["slug"] == "best-cake"
assert revision["status"] == "open"
assert revision["workspace_id"] == 2
assert revision["revision_id"] == 8
assert revision["sub_content_types"]
assert revision["revision_type"] == "creation"
assert revision["comment_ids"] == [18, 19, 20]
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 1
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Global manager"
assert revision["file_extension"] == ".thread.html"
assert revision["filename"] == "Best Cake.thread.html"
revision = revisions[1]
assert revision["content_type"] == "thread"
assert revision["content_id"] == 7
assert revision["is_archived"] is False
assert revision["is_deleted"] is False
assert revision["is_editable"] is True
assert revision["label"] == "Best Cakes?"
assert revision["parent_id"] == 3
assert revision["show_in_ui"] is True
assert revision["slug"] == "best-cakes"
assert revision["status"] == "open"
assert revision["workspace_id"] == 2
assert revision["revision_id"] == 26
assert revision["revision_type"] == "edition"
assert revision["sub_content_types"]
assert revision["comment_ids"] == []
# TODO - G.M - 2018-06-173 - check date format
assert revision["created"]
assert revision["author"]
assert revision["author"]["user_id"] == 3
assert revision["author"]["avatar_url"] is None
assert revision["author"]["public_name"] == "Bob i."
assert revision["file_extension"] == ".thread.html"
assert revision["filename"] == "Best Cakes?.thread.html"
def test_api__get_thread_revisions__ok_200__most_revision_type(self) -> None:
"""
get threads revisions
"""
dbsession = get_tm_session(self.session_factory, transaction.manager)
admin = dbsession.query(User).filter(User.email == "admin@admin.admin").one()
workspace_api = WorkspaceApi(current_user=admin, session=dbsession, config=self.app_config)
business_workspace = workspace_api.get_one(1)
content_api = ContentApi(current_user=admin, session=dbsession, config=self.app_config)
tool_folder = content_api.get_one(1, content_type=content_type_list.Any_SLUG)
test_thread = content_api.create(
content_type_slug=content_type_list.Thread.slug,
workspace=business_workspace,
parent=tool_folder,
label="Test Thread",
do_save=True,
do_notify=False,
)
with new_revision(session=dbsession, tm=transaction.manager, content=test_thread):
content_api.update_content(
test_thread, new_label="test_thread_updated", new_content="Just a test"
)
content_api.save(test_thread)
with new_revision(session=dbsession, tm=transaction.manager, content=test_thread):
content_api.archive(test_thread)
content_api.save(test_thread)
with new_revision(session=dbsession, tm=transaction.manager, content=test_thread):
content_api.unarchive(test_thread)
content_api.save(test_thread)
with new_revision(session=dbsession, tm=transaction.manager, content=test_thread):
content_api.delete(test_thread)
content_api.save(test_thread)
with new_revision(session=dbsession, tm=transaction.manager, content=test_thread):
content_api.undelete(test_thread)
content_api.save(test_thread)
dbsession.flush()
transaction.commit()
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
res = self.testapp.get(
"/api/v2/workspaces/1/threads/{}/revisions".format(test_thread.content_id), status=200
)
revisions = res.json_body
assert len(revisions) == 6
for revision in revisions:
assert revision["content_type"] == "thread"
assert revision["workspace_id"] == 1
assert revision["content_id"] == test_thread.content_id
revision = revisions[0]
assert revision["revision_type"] == "creation"
assert revision["is_editable"] is False
revision = revisions[1]
assert revision["revision_type"] == "edition"
assert revision["is_editable"] is False
revision = revisions[2]
assert revision["revision_type"] == "archiving"
assert revision["is_editable"] is False
revision = revisions[3]
assert revision["revision_type"] == "unarchiving"
assert revision["is_editable"] is False
revision = revisions[4]
assert revision["revision_type"] == "deletion"
assert revision["is_editable"] is False
revision = revisions[5]
assert revision["revision_type"] == "undeletion"
assert revision["is_editable"] is True
def test_api__set_thread_status__ok_200__nominal_case(self) -> None:
"""
Set thread status
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "closed-deprecated"}
# before
res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200)
content = res.json_body
assert content["content_type"] == "thread"
assert content["content_id"] == 7
assert content["status"] == "open"
assert content["is_editable"] is True
# set status
self.testapp.put_json("/api/v2/workspaces/2/threads/7/status", params=params, status=204)
# after
res = self.testapp.get("/api/v2/workspaces/2/threads/7", status=200)
content = res.json_body
assert content["content_type"] == "thread"
assert content["content_id"] == 7
assert content["status"] == "closed-deprecated"
assert content["is_editable"] is False
def test_api__set_thread_status__ok_400__wrong_status(self) -> None:
"""
Set thread status
"""
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "unexistant-status"}
res = self.testapp.put_json(
"/api/v2/workspaces/2/threads/7/status", params=params, status=400
)
# INFO - G.M - 2018-09-10 - Handle by marshmallow schema
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.GENERIC_SCHEMA_VALIDATION_ERROR
def test_api__set_thread_status__err_400__same_status(self) -> None:
self.testapp.authorization = ("Basic", ("admin@admin.admin", "admin@admin.admin"))
params = {"status": "open"}
res = self.testapp.put_json(
"/api/v2/workspaces/2/threads/7/status", params=params, status=400
)
assert res.json_body
assert "code" in res.json_body
assert res.json_body["code"] == ErrorCode.INVALID_STATUS_CHANGE
| 46.768891 | 116 | 0.631496 | 22,803 | 190,630 | 5.0396 | 0.014998 | 0.05778 | 0.064089 | 0.055344 | 0.978219 | 0.971884 | 0.967829 | 0.958231 | 0.947667 | 0.938869 | 0 | 0.017034 | 0.245517 | 190,630 | 4,075 | 117 | 46.780368 | 0.781967 | 0.040088 | 0 | 0.841667 | 0 | 0.003161 | 0.17451 | 0.039783 | 0 | 0 | 0 | 0.005644 | 0.327874 | 1 | 0.030172 | false | 0 | 0.005172 | 0 | 0.037644 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
7e47b4d7ceb33445c32cfe6565caf6173d824aa9 | 169 | py | Python | textattack/attack_results/__init__.py | cclauss/TextAttack | 98b8d6102aa47bf3c41afedace0215d48f8ed046 | [
"MIT"
] | 1 | 2021-06-24T19:35:18.000Z | 2021-06-24T19:35:18.000Z | textattack/attack_results/__init__.py | 53X/TextAttack | e6a7969abc1e28a2a8a7e2ace709b78eb9dc94be | [
"MIT"
] | null | null | null | textattack/attack_results/__init__.py | 53X/TextAttack | e6a7969abc1e28a2a8a7e2ace709b78eb9dc94be | [
"MIT"
] | 1 | 2021-11-12T05:26:21.000Z | 2021-11-12T05:26:21.000Z | from .failed_attack_result import FailedAttackResult
from .skipped_attack_result import SkippedAttackResult
from .successful_attack_result import SuccessfulAttackResult
| 42.25 | 60 | 0.911243 | 18 | 169 | 8.222222 | 0.555556 | 0.243243 | 0.364865 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.071006 | 169 | 3 | 61 | 56.333333 | 0.942675 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 7 |
7e6895509707dd5ac5d1572800cdb46aafc739c4 | 2,352 | py | Python | Evaluation/cluster.py | wangaxe/kmeans_defense | 2dc0ac9aa7fea1bffcacdbf3f0f02694f3b6a328 | [
"MIT"
] | null | null | null | Evaluation/cluster.py | wangaxe/kmeans_defense | 2dc0ac9aa7fea1bffcacdbf3f0f02694f3b6a328 | [
"MIT"
] | null | null | null | Evaluation/cluster.py | wangaxe/kmeans_defense | 2dc0ac9aa7fea1bffcacdbf3f0f02694f3b6a328 | [
"MIT"
] | null | null | null | from sklearn.cluster import KMeans, MiniBatchKMeans
import numpy as np
import torch
def kmean_process(original_img, n_cluster_center):
depth, height, width = original_img.shape
pixel_sample = np.reshape(original_img, (height*width, depth))
compressed_img = np.zeros((height, width, depth), dtype=np.float32)
kmeans = KMeans(n_clusters=n_cluster_center)
kmeans.fit(pixel_sample)
cluster_assignments = kmeans.predict(pixel_sample)
cluster_centers = kmeans.cluster_centers_
pixel_count = 0
for i in range(height):
for j in range(width):
cluster_idx = cluster_assignments[pixel_count]
cluster_value = cluster_centers[cluster_idx]
compressed_img[i][j] = cluster_value
pixel_count += 1
compressed_img = np.reshape(compressed_img, (depth, height, width))
return compressed_img
def mini_bench_kmean_process(original_img, n_cluster_center, batch_size=100):
depth, height, width = original_img.shape
pixel_sample = np.reshape(original_img, (height*width, depth))
compressed_img = np.zeros((height, width, depth), dtype=np.float32)
kmeans = MiniBatchKMeans(n_clusters=n_cluster_center, batch_size=batch_size)
kmeans.fit(pixel_sample)
cluster_assignments = kmeans.predict(pixel_sample)
cluster_centers = kmeans.cluster_centers_
pixel_count = 0
for i in range(height):
for j in range(width):
cluster_idx = cluster_assignments[pixel_count]
cluster_value = cluster_centers[cluster_idx]
compressed_img[i][j] = cluster_value
pixel_count += 1
compressed_img = np.reshape(compressed_img, (depth, height, width))
return compressed_img
def Kmeans_cluster(in_tensor, k = 2):
examples = in_tensor.detach().cpu().numpy()
assert len(examples.shape) == 4
inter_res = np.array([kmean_process(example, k) for example in examples])
assert inter_res.shape == examples.shape
res = torch.from_numpy(inter_res).cuda()
return res
def mb_Kmeans_cluster(in_tensor, k = 2):
examples = in_tensor.detach().cpu().numpy()
assert len(examples.shape) == 4
inter_res = np.array([mini_bench_kmean_process(example, k) for example in examples])
assert inter_res.shape == examples.shape
res = torch.from_numpy(inter_res).cuda()
return res
| 36.184615 | 88 | 0.711735 | 316 | 2,352 | 5.031646 | 0.202532 | 0.081761 | 0.03522 | 0.028931 | 0.918239 | 0.877987 | 0.877987 | 0.831447 | 0.831447 | 0.831447 | 0 | 0.007895 | 0.192177 | 2,352 | 64 | 89 | 36.75 | 0.828947 | 0 | 0 | 0.784314 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.078431 | 1 | 0.078431 | false | 0 | 0.058824 | 0 | 0.215686 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
7e811ebe6327b53430261c8c363fb3b3fb53953d | 44,984 | py | Python | py/genetargeter/recodeSelection.py | pablocarderam/genetargeter | 4cd257d02e0753c1be9fa339aa505bb895315a6d | [
"MIT"
] | null | null | null | py/genetargeter/recodeSelection.py | pablocarderam/genetargeter | 4cd257d02e0753c1be9fa339aa505bb895315a6d | [
"MIT"
] | 2 | 2021-06-02T15:28:52.000Z | 2021-06-17T12:08:53.000Z | py/genetargeter/recodeSelection.py | pablocarderam/genetargeter | 4cd257d02e0753c1be9fa339aa505bb895315a6d | [
"MIT"
] | null | null | null |
from builtins import str
from builtins import range
from py.utils.BioUtils import *; # Imports utils
from py.utils.GenBankToolbox import *; # Imports utils
from py.genetargeter.constants import *; # Imports constants
from gRNAScores.gRNAScoring import *; # Import scoring metrics
"""
Chooses the region to be recoded to avoid gRNA targeting in already transfected
regions. Returns GenBankAnn object with recoded sequence and indexes between
which it should go. GenBank object given as argument should contain one gene
with geneName included in its label, and at least one annotation with "LHR" in
its label. Also needs all gRNAs to be annotated in the file. Returns empty
region if LHR end is at or downstream of gene stop codon. Checks against
restriction sites given as parameters. Checks that gRNA recoded sequence has a
pairwise off-target score lower than the given threshold with respect to the
original gRNA.
"""
def chooseRecodeRegion3Prime(geneGB, gene, offTargetMethod="cfd", pamType="NGG", orgCodonTable=codonUsage(), targetRegionOverride=False, filterCutSites=[cut_FseI,cut_AsiSI,cut_IPpoI,cut_ISceI,cut_AflII,cut_AhdI,cut_BsiWI,cut_NheI], codonSampling=False, offScoreThreshold=10, minGCEnd=0.375, gRNATableString=""):
#TODO: debug #TODO: Recoded if upstream of stop codon add recode values to table
gRNAs = geneGB.findAnnsLabel("gRNA", True); # List of all gRNAs
gRNATable = gRNATableString.split('\n'); # split string into lines
gRNATable = [g.split(',') for g in gRNATable]; # split each line into values
if offTargetMethod == "hsu": # if off-target scoring with Hsu scores
offScoreThreshold = 1; # set threshold to 1%
log = ""; # init log
LHRs = geneGB.findAnnsLabel("LHR"); # LHR annotation objects
RHRs = geneGB.findAnnsLabel("RHR"); # LHR annotation objects
LHR = None
RHR = None
for ann in LHRs:
if '(unused)' not in ann.label:
LHR = ann
break
for ann in RHRs:
if '(unused)' not in ann.label:
RHR = ann
break
annRecoded = GenBankAnn(); # creates GenBankAnn object to hold recoded region
if LHR.index[1] < gene.index[1]: # if end of LHR is inside gene (or before)
startRecode = max(LHR.index[1], gene.index[0]); # start of recode region (end of LHR or start of gene, most downstream)
while not geneGB.checkInExon(startRecode) and startRecode <= len(geneGB.origin): # while recode region start is in intron,
startRecode += 1; # shift downstream
intronStartIndices = []; # stores start indexes of introns starting after recode sequence start
intronEndIndices = []; # stores end indexes of introns starting after recode sequence start
for ann in geneGB.findAnnsLabel(gene.label): # loop through annotations associated with transcript
if ann.type == "CDS": # if annotation is cds
if gene.index[1] > ann.index[1] > startRecode: # if annotation is an exon ending before gene end and after recode start,
intronStartIndices.append(ann.index[1]); # add this intron start index
if gene.index[1] > ann.index[0] > startRecode: # if annotation is an exon starting after recode start,
intronEndIndices.append(ann.index[0]); # add this intron end index
# if len(intronStartIndices) == 0: # if no CDS exons found in this way,
# for ann in geneGB.findAnnsLabel(gene.label.split('.')[0]): # loop through annotations associated with gene
# if ann.type == "exon": # if annotation is exon
# if gene.index[1] > ann.index[1] > startRecode: # if annotation is an exon ending before gene end and after recode start,
# intronStartIndices.append(ann.index[1]); # add this intron start index
# if gene.index[1] > ann.index[0] > startRecode: # if annotation is an exon starting after recode start,
# intronEndIndices.append(ann.index[0]); # add this intron end index
#
#
# elif ann.type == "intron": # if annotation is intron,
# if ann.index[0] > startRecode: # if annotation is an intron starting after recode start,
# intronStartIndices.append(ann.index[0]); # add this intron start index
# intronEndIndices.append(ann.index[1]); # add this intron start index
intronIndices = []; # will contain final indexes of introns downstream of recode start (introns to be removed from recoded region)
intronStartIndices = sorted(intronStartIndices); # sort
intronEndIndices = sorted(intronEndIndices); # sort
for i in range(len(intronEndIndices)): # for every intron end,
if intronEndIndices[i] >= LHR.index[1]: # if after LHR end,
start = 0; # will store corresponding intron start (largest underneath intron end)
for startIndex in intronStartIndices: # loop through starts
if startIndex > intronEndIndices[i]: # if start surpasses this end,
break; # stop loop
else: # if not,
start = startIndex; # set as start
intronIndices.append([start,intronEndIndices[i]]); # add these coordinates to intron splice list
#for g in gRNAs: # for every gRNA,
#if not (intronStartIndices[i] >= g.index[1] or intronEndIndices[i] <= g.index[0]): # if intron and gRNA overlap,
#intronIndices.append([intronStartIndices[i],intronEndIndices[i]]); # add these coordinates to intron splice list
endRecode = min(gene.index[1],RHR.index[0]) if targetRegionOverride else gene.index[1] - 3; # end of recode region (end of gene, exclude stop codon)
recodeSeq = geneGB.origin[startRecode:endRecode]; # will contain sequence to be recorded
nonRecodedEnd = ''
frame2 = 0
if len(intronIndices) > 0 and intronIndices[0][0] < endRecode: # if there are introns inside the target region,
recodeSeq = geneGB.origin[startRecode:intronIndices[0][0]]; # get recode sequence until first intron
for i in range(len(intronIndices)-1): # for every intron except last one,
if intronIndices[i][1] < endRecode:
recodeSeq = recodeSeq + geneGB.origin[intronIndices[i][1]:min(intronIndices[i+1][0],endRecode)]; # add next exon to recode seq
if intronIndices[len(intronIndices)-1][1] < endRecode:
recodeSeq = recodeSeq + geneGB.origin[intronIndices[len(intronIndices)-1][1]:endRecode]; # get rest of recode sequence until endRecode
# Adjust frame if not recoding to the stop codon
if targetRegionOverride:
restSeq = geneGB.origin[startRecode:intronIndices[0][0]]; # get recode sequence until first intron
for i in range(len(intronIndices)-1): # for every intron except last one,
restSeq = restSeq + geneGB.origin[intronIndices[i][1]:intronIndices[i+1][0]]; # add next exon to recode seq
restSeq = restSeq + geneGB.origin[intronIndices[len(intronIndices)-1][1]:gene.index[1]]; # get rest of recode sequence until endRecode
frame2 = 3-((len(restSeq)-len(recodeSeq)) % 3); # stores reading frame, index from start of sequence to be recoded
frame2 = frame2 if frame2 != 3 else 0
endRecode -= frame2; # modify recode start site according to reading frame
nonRecodedEnd = recodeSeq[-frame2:] if frame2!=0 else ''; # stores 0, 1 or 2 nucleotides not recoded due to reading frame
recodeSeq = recodeSeq[0:len(recodeSeq)-frame2]; # adjust recode region
frame = len(recodeSeq) % 3; # stores reading frame, index from start of sequence to be recoded
startRecode += frame; # modify recode start site according to reading frame
nonRecodedStart = recodeSeq[0:frame]; # stores 0, 1 or 2 nucleotides not recoded due to reading frame
recodeSeq = recodeSeq[frame:len(recodeSeq)]; # adjust recode region
cutSeqs = filterCutSites + [g.seq for g in gRNAs]; # list of all cut seqs. all gRNAs in gene are to be included as cut sequences
cutCheck = True; # variable used to check if no cut sequences are present.
offScore = 100; # stores off-target score. Default is 100% due to the fact that gRNA sequence is the same.
count = 0; # iteration counter
recodedSeq = recodeSeq; # assign recoded sequence to same as original
bestRecodedSeq = recodedSeq; # will store best candidate sequence
if len(recodeSeq) > 2: # if recodeSeq contains at least one codon,
tricky = -1; # > -1 if suspected to be hard to synthesize
badStart = False; # True if first bases have low melting temp (important for Gibson assembly)
candidateFound = False; # signal possible candidate found
bestRecodedSeq = recodedSeq; # will store best candidate sequence
while not cutCheck or offScore > offScoreThreshold or tricky > -1 or badStart: # while cutCheck shows hits in a cut sequences, or while the pairwise off-target score is over the threshold, or while there are difficult-to-synthesize structures in the recoded region, or while the first 40 bp have a bad gc content
if count == 1: # if recoded region has failed checks once,
log = log + "Defaulted recoded region recodonization to codon sampling due to possible difficulties in synthesis or enzyme cut sites.\n\n"; # log warning
codonSampling = True; # forces codonSampling to true
if count == 10:
log = log + "Defaulted recoded region recodonization to random codon sampling due to possible difficulties in synthesis or enzyme cut sites.\n\n"; # log warning
orgCodonTable = codonUsage(); # forces random codon selection to true
cutCheck = True; # reset cutCheck
offScore = 0; # reset offScore
tricky = -1; # reset tricky index
badStart = False; # reset badStart Boolean
recodedSeq = optimizeCodons(recodeSeq,orgCodonTable,codonSampling=codonSampling); # optimize codons.
for g in gRNAs: # for every gRNA candidate within recoded region,
if g.index[0] >= startRecode-frame and g.index[1] <= endRecode+frame2: # if grna is inside recoded region
gOnSeq = g.seq; # get original gRNA sequence
wholeRecSeq = nonRecodedStart + recodedSeq; # add initial bases
gOffSeq = "";
anchor = -1; # will store index of gRNA bp most to the left (whichever strand). Default to -1 to indicate excision
if geneGB.checkInExon(g.index[0]) or geneGB.checkInExon(g.index[1]): # if the gRNA hasn't been completely excised,
if pamType == "NGG" and g.comp or pamType == "TTTV" and not g.comp: # if PAM is to the left of the rest of the gRNA sequence (on whichever strand),
anchor = g.index[0]-startRecode-frame; # stores index of gRNA bp most to the left (whichever strand)
for intron in intronIndices: # for every intron,
if g.index[0] > intron[1]: # if anchor after end of intron,
anchor -= intron[1]-intron[0]; # substract intron length from anchor index
elif intron[0] >= g.index[0] >= intron[1]: # if anchor inside intron,
anchor -= g.index[0] - intron[0]; # substract distance between intron start and anchor from anchor
gOffSeq = wholeRecSeq[anchor:anchor+len(g.seq)]; # get recoded sequence that used to be gRNA
if g.comp: # if on comp strand
gOffSeq = revComp(gOffSeq); # save as reverse complement
else: # if PAM is to the right,
anchor = g.index[1]-startRecode-frame; # stores index of gRNA bp most to the right (whichever strand)
for intron in intronIndices: # for every intron,
if g.index[1] > intron[1]: # if anchor after end of intron,
anchor -= intron[1]-intron[0]; # substract intron length from anchor index
elif intron[0] >= g.index[1] >= intron[1]: # if anchor inside intron,
anchor -= g.index[1] - intron[0]; # substract distance between intron start and anchor from anchor
gOffSeq = wholeRecSeq[anchor-len(g.seq):anchor]; # get recoded sequence that used to be gRNA
if g.comp: # if on comp strand
gOffSeq = revComp(gOffSeq); # save as reverse complement
gNewPAM = ""; # will store new PAM sequence
if pamType == "NGG" and anchor > -1: # if using NGG PAM and gRNA not completely excised,
if (g.index[1]+3 >= endRecode and not g.comp) or (g.index[0]-3 >= startRecode and g.comp): # if PAM is within recoded region,
if not g.comp: # if on positive strand,
gNewPAM = wholeRecSeq[anchor+len(g.seq):anchor+len(g.seq)+3]; # retrieve PAM downstream of gRNA sequence
else: # if on negative strand,
gNewPAM = revComp(wholeRecSeq[anchor+len(g.seq)-3:anchor+len(g.seq)]); # retrieve PAM upstream of gRNA sequence, on comp strand
else: # if outside recoded region,
if g.comp: # if on comp strand,
gNewPAM = geneGB.origin[g.index[1]:g.index[1]+3]; # will store new PAM sequence
else: # if on positive strand,
gNewPAM = revComp(geneGB.origin[g.index[0]-3:g.index[0]]); # will store new PAM sequence
elif pamType == "TTTV" and anchor > -1: # if using TTTV PAM and gRNA not completely excised,
if (g.index[1]+4 >= endRecode and g.comp) or (g.index[0]-4 >= startRecode and not g.comp): # if PAM is inside recoded region,
if not g.comp: # if on positive strand,
gNewPAM = wholeRecSeq[anchor+len(g.seq)-4:anchor+len(g.seq)]; # retrieve PAM upstream of gRNA sequence
else: # if on negative strand,
gNewPAM = revComp(wholeRecSeq[anchor+len(g.seq):anchor+len(g.seq)+4]); # retrieve PAM downstream of gRNA sequence, on comp strand
else: # if outside recoded region,
if g.comp: # if on comp strand,
gNewPAM = geneGB.origin[g.index[1]:g.index[1]+4]; # will store new PAM sequence
else: # if on positive strand,
gNewPAM = revComp(geneGB.origin[g.index[0]-4:g.index[0]]); # will store new PAM sequence
newOffScore = 0; # Assume gRNA was excised
if offTargetMethod == "cfd" and len(gOffSeq) > 22: # if using cfd and gRNA not completely excised,
newOffScore = pairScoreCFD(gOnSeq,gOffSeq,gNewPAM,pamType); # calculate pairwise off-target score
elif offTargetMethod == "hsu" and len(gOffSeq) > 22: # if using hsu and gRNA not completely excised,
newOffScore = pairScoreHsu(gOnSeq,gOffSeq,gNewPAM,pamType); # calculate pairwise off-target score
offScore = max(offScore,newOffScore); # set offscore for next iteration
for g in gRNATable: # find this gRNA in table
if "gRNAs not evaluated" not in gRNATableString and g[14] == gOnSeq: # if there is a gRNA table (no table if using custom gRNA) and gRNA found,
g[15] = gOffSeq; # store recoded sequence
g[16] = str(newOffScore); # store recoded sequence's pair score
else: # if gRNA is not entirely contained,
offScore = max(offScore,0); # assume recoded
for site in cutSeqs: # for every cut site being filtered,
cutCheck = cutCheck * ( findFirst(recodedSeq,site) < 0 ); # Find cut site, register in cutCheck
cutCheck = cutCheck * ( findFirst(recodedSeq,revComp(site)) < 0 ); # Find cut site in comp strand, register in cutCheck
if gcContent(recodedSeq[0:40]) < minGCEnd: # if the first bases don't have enough gc content
badStart = True;
trickyCount = 1
trickyLimit = 1000
tricky = isTricky(recodedSeq); # check if tricky to synthesize
bestRecodedSeq = recodedSeq if bestRecodedSeq==recodeSeq else bestRecodedSeq; # store this sequence if no recoded sequence has been stored as best
if offScore <= offScoreThreshold and cutCheck: # if parameters other than badStart are ok and this sequence has better start than previous best,
if not candidateFound or isTricky(bestRecodedSeq) > -1: # if no candidate found until now or current best is already tricky,
while tricky > -1 and tricky < len(recodedSeq)-9 and trickyCount < trickyLimit: # targeted recoding of problematic fragments
recodedSeq = recodedSeq[0:tricky-tricky%3] + optimizeCodons(recodedSeq[tricky-tricky%3:tricky-tricky%3+9]) + recodedSeq[tricky-tricky%3+9:]; # optimize codons.
new_tricky = isTricky(recodedSeq)
tricky = max(tricky,new_tricky) if new_tricky > -1 else new_tricky; # check if tricky to synthesize (only downstream to avoid going back to fix newly repeated sequences)
trickyCount += 1
if trickyCount % 10 == 0: # shuffle everything every 100 targeted recodings
recodedSeq = recodedSeq[0:tricky-tricky%3] + optimizeCodons(recodedSeq[tricky-tricky%3:]); # optimize codons of remainder
new_tricky = isTricky(recodedSeq)
tricky = max(tricky,new_tricky) if new_tricky > -1 else new_tricky; # check if tricky to synthesize (only downstream to avoid going back to fix newly repeated sequences)
bestRecodedSeq = recodedSeq; # make this new best
elif not tricky > -1 and gcContent(recodedSeq[0:40]) > gcContent(bestRecodedSeq[0:40]):
bestRecodedSeq = recodedSeq; # make this new best
if not tricky > -1:
candidateFound = True; # signal possible candidate found
count += 1; # advances iteration counter
if count > 200 or trickyCount >= trickyLimit: # if out of iteration limit,
if not candidateFound: # if no candidate without cut sequences found,
if tricky > -1:
log = log + "Warning: Recoded region for gene " + gene.label + " could not reshuffle enough to avoid repeated sequences or low-complexity regions.\n\n"; # log warning
else:
log = log + "Warning: Recoded region for gene " + gene.label + " could not reshuffle enough to fulfill the maximum off-target sgRNA score threshold, or avoid all the following cut sequences: \n" + str(cutSeqs) + "\n\n"; # log warning
break; # escape loop
#print [gOnSeq+"NGG",gOffSeq+gNewPAM,pairScoreCFD(gOnSeq,gOffSeq,gNewPAM,pamType),pairScoreHsu(gOnSeq,gOffSeq,gNewPAM,pamType)]
recodedSeq = nonRecodedStart + bestRecodedSeq + nonRecodedEnd; # adds initial bases from reading frame adjustment to best candidate
annRecoded = GenBankAnn(gene.label + " Recoded", "misc_feature", recodedSeq, False, [startRecode-frame,endRecode+frame2], annColors['recodedRegionColor']); # creates var to store finished recodedSeq as annotation
log = log + "Recoded region with size " + str(len(recodedSeq)) + " for gene " + gene.label + " selected.\n\n"; # logs this process finished
else: # if no recoded region necessary,
log = log + "Recoded region not deemed necessary for gene " + gene.label + ".\n\n"; # logs this process finished
if "gRNAs not evaluated" not in gRNATableString:
gRNATableString = "\n".join([",".join(g) for g in gRNATable]); # Creates string from grna array
gRNATableString = gRNATableString.replace(">=threshold",">="+str(offScoreThreshold)); # adds pairwise recoded threshold values
return {"out":annRecoded, "log":log, "gRNATable":gRNATableString}; # returns recoded region GenBankAnn object
"""
Chooses the region to be recoded to avoid gRNA targeting in already transfected
regions. Returns GenBankAnn object with recoded sequence and indexes between
which it should go. GenBank object given as argument should contain one gene
with geneName included in its label, and at least one annotation with "LHR" in
its label. Also needs all gRNAs to be annotated in the file. Returns empty
region if LHR end is at or downstream of gene stop codon. Checks against
restriction sites given as parameters. Checks that gRNA recoded sequence has a
pairwise off-target score lower than the given threshold with respect to the
original gRNA.
"""
def chooseRecodeRegion5Prime(geneGB, gene, offTargetMethod="cfd", pamType="NGG", orgCodonTable=codonUsage(), targetRegionOverride=False, filterCutSites=[cut_FseI,cut_AsiSI,cut_IPpoI,cut_ISceI,cut_AflII,cut_AhdI,cut_BsiWI,cut_NheI], codonSampling=False, offScoreThreshold=10, minGCEnd=0.375, gRNATableString="", haTag=True):
#TODO: debug #TODO: Recoded if upstream of stop codon add recode values to table
gRNAs = geneGB.findAnnsLabel("gRNA", True); # List of all gRNAs
gRNATable = gRNATableString.split('\n'); # split string into lines
gRNATable = [g.split(',') for g in gRNATable]; # split each line into values
if offTargetMethod == "hsu": # if off-target scoring with Hsu scores
offScoreThreshold = 1; # set threshold to 1%
log = ""; # init log
LHR = geneGB.findAnnsLabel("LHR")[0]; # LHR annotation object
RHR = geneGB.findAnnsLabel("RHR")[0]; # RHR annotation object
annRecoded = GenBankAnn(); # creates GenBankAnn object to hold recoded region
if RHR.index[0] > gene.index[0]: # if end of RHR is inside gene
endRecode = min(RHR.index[0],gene.index[1]); # end of recode region (start of RHR or end of gene, most upstream)
while not geneGB.checkInExon(endRecode): # while recode region end is in intron,
endRecode -= 1; # shift upstream
intronStartIndices = []; # stores start indexes of introns starting after recode sequence start
intronEndIndices = []; # stores end indexes of introns starting after recode sequence start
for ann in geneGB.findAnnsLabel(gene.label): # loop through annotations associated with transcript
if ann.type == "CDS": # if annotation is cds
if gene.index[0] < ann.index[1] < endRecode: # if annotation is an exon ending after gene start and before recode end,
intronStartIndices.append(ann.index[1]); # add this intron start index
if gene.index[0] < ann.index[0] < endRecode: # if annotation is an exon starting before recode end,
intronEndIndices.append(ann.index[0]); # add this intron end index
# if len(intronStartIndices) == 0: # if no CDS exons found in this way,
# for ann in geneGB.findAnnsLabel(gene.label.split('.')[0]): # loop through annotations associated with gene
# if ann.type == "exon": # if annotation is exon
# if gene.index[0] < ann.index[1] < endRecode: # if annotation is an exon ending after gene start and before recode end,
# intronStartIndices.append(ann.index[1]); # add this intron start index
# if gene.index[0] < ann.index[0] < endRecode: # if annotation is an exon starting before recode end,
# intronEndIndices.append(ann.index[0]); # add this intron end index
#
#
# elif ann.type == "intron": # if annotation is intron,
# if ann.index[1] < endRecode: # if annotation is an intron ending before recode end,
# intronStartIndices.append(ann.index[0]); # add this intron start index
# intronEndIndices.append(ann.index[1]); # add this intron start index
intronIndices = []; # will contain final indexes of introns downstream of recode start (introns to be removed from recoded region)
intronStartIndices = sorted(intronStartIndices); # sort
intronEndIndices = sorted(intronEndIndices); # sort
for i in range(len(intronEndIndices)): # for every intron end,
if intronEndIndices[i] <= RHR.index[0]: # if before RHR start,
start = 0; # will store corresponding intron start (largest underneath intron end)
for startIndex in intronStartIndices: # loop through starts
if startIndex > intronEndIndices[i]: # if start surpasses this end,
break; # stop loop
else: # if not,
start = startIndex; # set as start
intronIndices.append([start,intronEndIndices[i]]); # add these coordinates to intron splice list
#for g in gRNAs: # for every gRNA,
#if not (intronStartIndices[i] >= g.index[1] or intronEndIndices[i] <= g.index[0]): # if intron and gRNA overlap,
#intronIndices.append([intronStartIndices[i],intronEndIndices[i]]); # add these coordinates to intron splice list
startRecode = max(gene.index[0],LHR.index[1]) if targetRegionOverride else gene.index[0]; # end of recode region (end of gene, exclude stop codon)
recodeSeq = geneGB.origin[startRecode:endRecode]; # will contain sequence to be recorded
nonRecodedStart = ''
frame2 = 0
if len(intronIndices) > 0 and intronIndices[0][0] < endRecode: # if there are introns inside the target region,
recodeSeq = geneGB.origin[startRecode:intronIndices[0][0]]; # get recode sequence until first intron
for i in range(len(intronIndices)-1): # for every intron except last one,
if intronIndices[i][1] < endRecode:
recodeSeq = recodeSeq + geneGB.origin[intronIndices[i][1]:min(intronIndices[i+1][0],endRecode)]; # add next exon to recode seq
if intronIndices[len(intronIndices)-1][1] < endRecode:
recodeSeq = recodeSeq + geneGB.origin[intronIndices[len(intronIndices)-1][1]:endRecode]; # get rest of recode sequence until endRecode
# Adjust frame if not recoding from the start codon
if targetRegionOverride:
restSeq = geneGB.origin[startRecode:intronIndices[0][0]]; # get recode sequence until first intron
for i in range(len(intronIndices)-1): # for every intron except last one,
restSeq = restSeq + geneGB.origin[intronIndices[i][1]:intronIndices[i+1][0]]; # add next exon to recode seq
restSeq = restSeq + geneGB.origin[intronIndices[len(intronIndices)-1][1]:gene.index[1]]; # get rest of recode sequence until endRecode
frame2 = len(recodeSeq) % 3; # stores reading frame, index from start of sequence to be recoded
frame = 3-((len(restSeq)-len(recodeSeq)) % 3) # stores reading frame, index from start of sequence to be recoded
frame = frame if frame != 3 else 0
startRecode += frame2; # modify recode start site according to reading frame
nonRecodedStart = recodeSeq[0:frame2] if frame2!=0 else ''; # stores 0, 1 or 2 nucleotides not recoded due to reading frame
recodeSeq = recodeSeq[frame2:]; # adjust recode region
else:
frame = len(recodeSeq) % 3; # stores reading frame, index from start of sequence to be recoded
# frame = len(recodeSeq) % 3; # stores reading frame, index from start of sequence to be recoded
endRecode -= frame; # modify recode end site according to reading frame
nonRecodedEnd = ""; # stores 0, 1 or 2 nucleotides not recoded due to reading frame
if frame != 0: # if frame shift is not zero, to avoid listing all of recode region with recodeSeq[-0:],
nonRecodedEnd = recodeSeq[-frame:]; # stores 0, 1 or 2 nucleotides not recoded due to reading frame
recodeSeq = recodeSeq[0:len(recodeSeq)-frame]; # adjust recode region
if haTag: # if adding an HA tag,
recodeSeq = ha_tag + recodeSeq; # add HA tag to start of recoded region
cutSeqs = filterCutSites + [g.seq for g in gRNAs]; # list of all cut seqs. all gRNAs in gene are to be included as cut sequences
cutCheck = True; # variable used to check if cut sequences are present. Initially false since all gRNAs are present.
offScore = 100; # stores off-target score. Default is 100% due to the fact that gRNA sequence is the same.
count = 0; # iteration counter
recodedSeq = recodeSeq; # assign recoded sequence to same as original
bestRecodedSeq = recodedSeq; # will store best candidate sequence
if len(recodeSeq) > 2: # if recodeSeq contains at least one codon,
tricky = -1; # True if suspected to be hard to synthesize
badStart = False; # True if first bases have low melting temp (important for Gibson assembly)
candidateFound = False; # signal possible candidate found
bestRecodedSeq = recodedSeq; # will store best candidate sequence
while not cutCheck or offScore > offScoreThreshold or tricky > -1 or badStart: # while cutCheck is greater than what you would expect for no hits in all cut sequences plus the gRNAs on both positive and comp strands, or while the pairwise off-target score is over the threshold, or while there are difficult-to-synthesize structures in the recoded region, or while the first 40 bp have a bad gc content
if count == 1: # if recoded region has failed checks once,
log = log + "Defaulted recoded region recodonization to codon sampling due to possible difficulties in synthesis or enzyme cut sites.\n\n"; # log warning
codonSampling = True; # forces codonSampling to true
if count == 10:
log = log + "Defaulted recoded region recodonization to random codon sampling due to possible difficulties in synthesis or enzyme cut sites.\n\n"; # log warning
orgCodonTable = codonUsage(); # forces random codon selection to true
cutCheck = True; # reset cutCheck
offScore = 0; # reset offScore
tricky = -1; # reset tricky index
badStart = False; # reset badStart Boolean
recodedSeq = optimizeCodons(recodeSeq,orgCodonTable,codonSampling=codonSampling); # optimize codons.
for g in gRNAs: # for every gRNA candidate within recoded region,
if g.index[0] >= startRecode-frame2 and g.index[1] <= endRecode+frame: # if grna is inside recoded region
gOnSeq = g.seq; # get original gRNA sequence
wholeRecSeq = nonRecodedStart + recodedSeq + nonRecodedEnd; # add initial bases
gOffSeq = "";
anchor = -1; # will store index of gRNA bp most to the left (whichever strand). Default to -1 to indicate excision
if geneGB.checkInExon(g.index[0]) or geneGB.checkInExon(g.index[1]): # if the gRNA hasn't been completely excised,
if pamType == "NGG" and g.comp or pamType == "TTTV" and not g.comp: # if PAM is to the left of the rest of the gRNA sequence (on whichever strand),
anchor = g.index[0]-startRecode-frame2; # stores index of gRNA bp most to the left (whichever strand)
for intron in intronIndices: # for every intron,
if g.index[0] > intron[1]: # if anchor after end of intron,
anchor -= intron[1]-intron[0]; # substract intron length from anchor index
elif intron[0] >= g.index[0] >= intron[1]: # if anchor inside intron,
anchor -= g.index[0] - intron[0]; # substract distance between intron start and anchor from anchor
gOffSeq = wholeRecSeq[anchor:anchor+len(g.seq)]; # get recoded sequence that used to be gRNA
if g.comp: # if on comp strand
gOffSeq = revComp(gOffSeq); # save as reverse complement
else: # if PAM is to the right,
anchor = g.index[1]-startRecode-frame2; # stores index of gRNA bp most to the right (whichever strand)
for intron in intronIndices: # for every intron,
if g.index[1] > intron[1]: # if anchor after end of intron,
anchor -= intron[1]-intron[0]; # substract intron length from anchor index
elif intron[0] >= g.index[1] >= intron[1]: # if anchor inside intron,
anchor -= g.index[1] - intron[0]; # substract distance between intron start and anchor from anchor
gOffSeq = wholeRecSeq[anchor-len(g.seq):anchor]; # get recoded sequence that used to be gRNA
if g.comp: # if on comp strand
gOffSeq = revComp(gOffSeq); # save as reverse complement
gNewPAM = ""; # will store new PAM sequence
if pamType == "NGG" and anchor > -1: # if using NGG PAM and gRNA not completely excised,
if (g.index[1]+3 >= endRecode and not g.comp) or (g.index[0]-3 >= startRecode and g.comp): # if PAM is within recoded region,
if not g.comp: # if on positive strand,
gNewPAM = wholeRecSeq[anchor+len(g.seq):anchor+len(g.seq)+3]; # retrieve PAM downstream of gRNA sequence
else: # if on negative strand,
gNewPAM = revComp(wholeRecSeq[anchor+len(g.seq)-3:anchor+len(g.seq)]); # retrieve PAM upstream of gRNA sequence, on comp strand
else: # if outside recoded region,
if g.comp: # if on comp strand,
gNewPAM = geneGB.origin[g.index[1]:g.index[1]+3]; # will store new PAM sequence
else: # if on positive strand,
gNewPAM = revComp(geneGB.origin[g.index[0]-3:g.index[0]]); # will store new PAM sequence
elif pamType == "TTTV" and anchor > -1: # if using TTTV PAM and gRNA not completely excised,
if (g.index[1]+4 >= endRecode and g.comp) or (g.index[0]-4 >= startRecode and not g.comp): # if PAM is inside recoded region,
if not g.comp: # if on positive strand,
gNewPAM = wholeRecSeq[anchor+len(g.seq)-4:anchor+len(g.seq)]; # retrieve PAM upstream of gRNA sequence
else: # if on negative strand,
gNewPAM = revComp(wholeRecSeq[anchor+len(g.seq):anchor+len(g.seq)+4]); # retrieve PAM downstream of gRNA sequence, on comp strand
else: # if outside recoded region,
if g.comp: # if on comp strand,
gNewPAM = geneGB.origin[g.index[1]:g.index[1]+4]; # will store new PAM sequence
else: # if on positive strand,
gNewPAM = revComp(geneGB.origin[g.index[0]-4:g.index[0]]); # will store new PAM sequence
newOffScore = 0; # Assume gRNA was excised
if offTargetMethod == "cfd" and len(gOffSeq) > 22: # if using cfd and gRNA not completely excised,
newOffScore = pairScoreCFD(gOnSeq,gOffSeq,gNewPAM,pamType); # calculate pairwise off-target score
elif offTargetMethod == "hsu" and len(gOffSeq) > 22: # if using hsu and gRNA not completely excised,
newOffScore = pairScoreHsu(gOnSeq,gOffSeq,gNewPAM,pamType); # calculate pairwise off-target score
offScore = max(offScore,newOffScore); # set offscore for next iteration
for g in gRNATable: # find this gRNA in table
if "gRNAs not evaluated" not in gRNATableString and g[14] == gOnSeq: # if there is a gRNA table (no table if using custom gRNA) and gRNA found,
g[15] = gOffSeq; # store recoded sequence
g[16] = str(newOffScore); # store recoded sequence's pair score
else: # if gRNA is not entirely contained,
offScore = max(offScore,0); # assume recoded
for site in cutSeqs: # for every cut site being filtered,
cutCheck = cutCheck * ( findFirst(recodedSeq,site) < 0 ); # Find cut site, register in cutCheck
cutCheck = cutCheck * ( findFirst(recodedSeq,revComp(site)) < 0 ); # Find cut site in comp strand, register in cutCheck
if gcContent(recodedSeq[-40:]) < minGCEnd: # if the last bases don't have enough gc content
badStart = True;
trickyCount = 1
trickyLimit = 1000
tricky = isTricky(recodedSeq); # check if tricky to synthesize
bestRecodedSeq = recodedSeq if bestRecodedSeq==recodeSeq else bestRecodedSeq; # store this sequence if no recoded sequence has been stored as best
if offScore <= offScoreThreshold and cutCheck: # if parameters other than badStart are ok and this sequence has better start than previous best,
if not candidateFound or isTricky(bestRecodedSeq) > -1: # if no candidate found until now or current best is already tricky,
while tricky > -1 and tricky < len(recodedSeq)-9 and trickyCount < trickyLimit: # targeted recoding of problematic fragments
recodedSeq = recodedSeq[0:tricky-tricky%3] + optimizeCodons(recodedSeq[tricky-tricky%3:tricky-tricky%3+9]) + recodedSeq[tricky-tricky%3+9:]; # optimize codons.
new_tricky = isTricky(recodedSeq)
tricky = max(tricky,new_tricky) if new_tricky > -1 else new_tricky; # check if tricky to synthesize (only downstream to avoid going back to fix newly repeated sequences)
trickyCount += 1
if trickyCount % 10 == 0: # shuffle everything every 100 targeted recodings
recodedSeq = recodedSeq[0:tricky-tricky%3] + optimizeCodons(recodedSeq[tricky-tricky%3:]); # optimize codons of remainder
new_tricky = isTricky(recodedSeq)
tricky = max(tricky,new_tricky) if new_tricky > -1 else new_tricky; # check if tricky to synthesize (only downstream to avoid going back to fix newly repeated sequences)
bestRecodedSeq = recodedSeq; # make this new best
elif not tricky > -1 and gcContent(recodedSeq[-40:]) > gcContent(bestRecodedSeq[-40:]):
bestRecodedSeq = recodedSeq; # make this new best
if not tricky > -1:
candidateFound = True; # signal possible candidate found
count += 1; # advances iteration counter
if count > 200 or trickyCount >= trickyLimit: # if out of iteration limit,
if not candidateFound: # if no candidate without cut sequences found,
if tricky > -1:
log = log + "Warning: Recoded region for gene " + gene.label + " could not reshuffle enough to avoid repeated sequences or low-complexity regions.\n\n"; # log warning
else:
log = log + "Warning: Recoded region for gene " + gene.label + " could not reshuffle enough to fulfill the maximum off-target sgRNA score threshold, or avoid all the following cut sequences: \n" + str(cutSeqs) + "\n\n"; # log warning
break; # escape loop
#print [gOnSeq+"NGG",gOffSeq+gNewPAM,pairScoreCFD(gOnSeq,gOffSeq,gNewPAM,pamType),pairScoreHsu(gOnSeq,gOffSeq,gNewPAM,pamType)]
recodedSeq = nonRecodedStart + bestRecodedSeq + nonRecodedEnd; # adds end bases from reading frame adjustment to best candidate
annRecoded = GenBankAnn(gene.label + " Recoded", "misc_feature", recodedSeq, False, [startRecode-frame2,endRecode+frame], annColors['recodedRegionColor']); # creates var to store finished recodedSeq as annotation
haTagMsg = ""; # used to output message
if haTag: # if using an HA tag,
haTagMsg = " with a recoded HA tag"; # msg modifier
log = log + "Recoded region with size " + str(len(recodedSeq)) + " for gene " + gene.label + haTagMsg + " selected.\n\n"; # logs this process finished
else: # if no recoded region necessary,
log = log + "Recoded region not deemed necessary for gene " + gene.label + ".\n\n"; # logs this process finished
if "gRNAs not evaluated" not in gRNATableString:
gRNATableString = "\n".join([",".join(g) for g in gRNATable]); # Creates string from grna array
gRNATableString = gRNATableString.replace(">=threshold",">="+str(offScoreThreshold)); # adds pairwise recoded threshold values
return {"out":annRecoded, "log":log, "gRNATable":gRNATableString}; # returns recoded region GenBankAnn object
"""
Chooses the region to be recoded to avoid gRNA targeting in already transfected
regions. Returns GenBankAnn object with recoded sequence and indexes between
which it should go. GenBank object given as argument should contain one gene
with geneName included in its label, and at least one annotation with "LHR" in
its label. Also needs all gRNAs to be annotated in the file. Returns empty
region if LHR end is at or downstream of gene stop codon. Checks against
restriction sites given as parameters. Checks that gRNA recoded sequence has a
pairwise off-target score lower than the given threshold with respect to the
original gRNA.
"""
def chooseRecodeRegion(geneGB, gene, offTargetMethod="cfd", pamType="NGG", orgCodonTable=codonUsage(), targetRegionOverride=False, filterCutSites=[cut_FseI,cut_AsiSI,cut_IPpoI,cut_ISceI,cut_AflII,cut_AhdI,cut_BsiWI,cut_NheI], codonSampling=False, offScoreThreshold=10, minGCEnd=0.375, gRNATableString="", target3Prime=True, haTag=False):
out = {}; # will contain method output
if target3Prime: # if targeting 3'
out = chooseRecodeRegion3Prime(geneGB, gene, offTargetMethod, pamType=pamType, orgCodonTable=orgCodonTable,codonSampling=codonSampling, gRNATableString=gRNATableString, targetRegionOverride=targetRegionOverride, filterCutSites=filterCutSites); # defines region to be recoded, returns recoded sequence
else: # if using pSN150,
out = chooseRecodeRegion5Prime(geneGB, gene, offTargetMethod, pamType=pamType, orgCodonTable=orgCodonTable,codonSampling=codonSampling, gRNATableString=gRNATableString, haTag=haTag, targetRegionOverride=targetRegionOverride, filterCutSites=filterCutSites); # defines region to be recoded, returns recoded sequence
return out;
| 73.744262 | 414 | 0.618042 | 5,418 | 44,984 | 5.123477 | 0.083056 | 0.01124 | 0.006556 | 0.009366 | 0.927879 | 0.916424 | 0.913001 | 0.911884 | 0.909939 | 0.907922 | 0 | 0.014241 | 0.302219 | 44,984 | 609 | 415 | 73.865353 | 0.870114 | 0.369131 | 0 | 0.80198 | 0 | 0.014851 | 0.06244 | 0 | 0 | 0 | 0 | 0.003284 | 0 | 1 | 0.007426 | false | 0 | 0.014851 | 0 | 0.029703 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
7e8a6ba7ba5315f5b907de2d512c604d0dcf914f | 88 | py | Python | pytestJune2020/code/fixtures/test_add_pytest.py | Parazyne/my_presentations | 0f4298761c05208ac2ab8f907cbd0471c60ad988 | [
"MIT"
] | 1 | 2020-06-19T03:15:22.000Z | 2020-06-19T03:15:22.000Z | pytestJune2020/code/simple/test_add_pytest.py | Parazyne/my_presentations | 0f4298761c05208ac2ab8f907cbd0471c60ad988 | [
"MIT"
] | null | null | null | pytestJune2020/code/simple/test_add_pytest.py | Parazyne/my_presentations | 0f4298761c05208ac2ab8f907cbd0471c60ad988 | [
"MIT"
] | null | null | null | from add import add_two_numbers
def test_add():
assert add_two_numbers(1, 3) == 4
| 14.666667 | 37 | 0.715909 | 16 | 88 | 3.625 | 0.6875 | 0.206897 | 0.448276 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.042254 | 0.193182 | 88 | 5 | 38 | 17.6 | 0.774648 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.333333 | 1 | 0.333333 | true | 0 | 0.333333 | 0 | 0.666667 | 0 | 1 | 0 | 0 | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 8 |
0e8368a3facf91ee997929f09d55d3a1d854d4b4 | 158 | py | Python | tests/cases/resources/tests/__init__.py | chop-dbhi/varify | 5dc721e49ed9bd3582f4b117785fdd1a8b6ba777 | [
"BSD-2-Clause"
] | 6 | 2015-01-16T14:35:29.000Z | 2017-06-18T05:56:15.000Z | tests/cases/resources/tests/__init__.py | solvebio/varify | 5dc721e49ed9bd3582f4b117785fdd1a8b6ba777 | [
"BSD-2-Clause"
] | null | null | null | tests/cases/resources/tests/__init__.py | solvebio/varify | 5dc721e49ed9bd3582f4b117785fdd1a8b6ba777 | [
"BSD-2-Clause"
] | 3 | 2015-05-27T15:03:17.000Z | 2020-03-11T08:42:46.000Z | from gene_rank import * # noqa
from assessments import * # noqa
from genes import * # noqa
from phenotypes import * # noqa
from variants import * # noqa
| 26.333333 | 33 | 0.71519 | 21 | 158 | 5.333333 | 0.428571 | 0.446429 | 0.5 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.221519 | 158 | 5 | 34 | 31.6 | 0.910569 | 0.151899 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 7 |
0e8f062a5146956f40087807fffcef101744fede | 2,560 | py | Python | tests/test_http_exception.py | VincentHch/fastapi-jsonrpc | aeba440322c30224bc121f5a5ce77ab354841989 | [
"MIT"
] | 155 | 2019-08-09T18:33:19.000Z | 2022-03-23T11:32:44.000Z | tests/test_http_exception.py | VincentHch/fastapi-jsonrpc | aeba440322c30224bc121f5a5ce77ab354841989 | [
"MIT"
] | 29 | 2020-01-16T11:34:08.000Z | 2022-02-16T15:23:57.000Z | tests/test_http_exception.py | VincentHch/fastapi-jsonrpc | aeba440322c30224bc121f5a5ce77ab354841989 | [
"MIT"
] | 24 | 2019-10-11T15:15:40.000Z | 2022-02-17T11:54:21.000Z | import contextlib
from json import dumps as json_dumps
from fastapi import HTTPException
from fastapi_jsonrpc import JsonRpcContext
def test_method(ep, raw_request):
@ep.method()
def probe() -> str:
raise HTTPException(401)
resp = raw_request(json_dumps({
'id': 1,
'jsonrpc': '2.0',
'method': 'probe',
'params': {},
}))
assert resp.status_code == 401
assert resp.json() == {'detail': 'Unauthorized'}
def test_ep_middleware_enter(ep, raw_request):
@contextlib.asynccontextmanager
async def middleware(_ctx: JsonRpcContext):
raise HTTPException(401)
# noinspection PyUnreachableCode
yield
ep.middlewares.append(middleware)
@ep.method()
def probe() -> str:
return 'qwe'
resp = raw_request(json_dumps({
'id': 1,
'jsonrpc': '2.0',
'method': 'probe',
'params': {},
}))
assert resp.status_code == 401
assert resp.json() == {'detail': 'Unauthorized'}
def test_ep_middleware_exit(ep, raw_request):
@contextlib.asynccontextmanager
async def middleware(_ctx: JsonRpcContext):
yield
raise HTTPException(401)
ep.middlewares.append(middleware)
@ep.method()
def probe() -> str:
return 'qwe'
resp = raw_request(json_dumps({
'id': 1,
'jsonrpc': '2.0',
'method': 'probe',
'params': {},
}))
assert resp.status_code == 401
assert resp.json() == {'detail': 'Unauthorized'}
def test_method_middleware_enter(ep, raw_request):
@contextlib.asynccontextmanager
async def middleware(_ctx: JsonRpcContext):
raise HTTPException(401)
# noinspection PyUnreachableCode
yield
@ep.method(middlewares=[middleware])
def probe() -> str:
return 'qwe'
resp = raw_request(json_dumps({
'id': 1,
'jsonrpc': '2.0',
'method': 'probe',
'params': {},
}))
assert resp.status_code == 401
assert resp.json() == {'detail': 'Unauthorized'}
def test_method_middleware_exit(ep, raw_request):
@contextlib.asynccontextmanager
async def middleware(_ctx: JsonRpcContext):
yield
raise HTTPException(401)
@ep.method(middlewares=[middleware])
def probe() -> str:
return 'qwe'
resp = raw_request(json_dumps({
'id': 1,
'jsonrpc': '2.0',
'method': 'probe',
'params': {},
}))
assert resp.status_code == 401
assert resp.json() == {'detail': 'Unauthorized'}
| 22.654867 | 52 | 0.598828 | 267 | 2,560 | 5.595506 | 0.164794 | 0.066934 | 0.040161 | 0.060241 | 0.892905 | 0.880187 | 0.880187 | 0.880187 | 0.880187 | 0.880187 | 0 | 0.023923 | 0.265234 | 2,560 | 112 | 53 | 22.857143 | 0.770335 | 0.023828 | 0 | 0.890244 | 0 | 0 | 0.098958 | 0 | 0 | 0 | 0 | 0 | 0.121951 | 1 | 0.121951 | false | 0 | 0.04878 | 0.04878 | 0.219512 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
0eb15ce5e2d074f96d46fea515a175efade9aeb4 | 9,220 | py | Python | icons/check_boxes.py | robert-hh/SSD1963-TFT-Library-for-PyBoard | db9786cdd95f9dab5334a9de28bed4e26436815c | [
"MIT"
] | 16 | 2016-02-23T12:20:36.000Z | 2021-02-02T06:41:49.000Z | icons/check_boxes.py | robert-hh/SSD1963-TFT-Library-for-PyBoard-and-RP2040 | db9786cdd95f9dab5334a9de28bed4e26436815c | [
"MIT"
] | 2 | 2016-11-26T07:46:58.000Z | 2017-12-10T08:44:38.000Z | icons/check_boxes.py | robert-hh/SSD1963-TFT-Library-for-PyBoard | db9786cdd95f9dab5334a9de28bed4e26436815c | [
"MIT"
] | 9 | 2016-06-04T08:22:55.000Z | 2020-04-19T14:40:36.000Z |
# Code generated by bmp_to_icon.py
from uctypes import addressof
_icons = {
0: (
b'\xff\xff\xff\xff\xff\xff\xff\xd7\xff\xff\xff\xff\xff\xff\xff\xd5'
b'\xff\xff\xff\xff\xff\xff\xff\x55\xc0\x00\x00\x00\x00\x00\xfd\x57'
b'\x00\x00\x00\x00\x00\x03\xf5\x57\x02\xaa\xaa\xaa\xaa\xaa\xa5\x5f'
b'\x02\xaa\xaa\xaa\xaa\xaa\x95\x7f\x02\xaa\xaa\xaa\xaa\xaa\x55\xff'
b'\x02\xaa\xaa\xaa\xaa\xaa\x55\xff\x02\xaa\xaa\xaa\xaa\xa9\x57\xcf'
b'\x02\xaa\xaa\xaa\xaa\xa5\x5b\x0f\x02\xaa\xaa\xaa\xaa\xa5\x58\x0f'
b'\x02\xaa\xaa\xaa\xaa\x95\x68\x0f\x02\xaa\xaa\xaa\xaa\x55\xa8\x0f'
b'\x02\xaa\x5a\xaa\xaa\x55\xa8\x0f\x02\xa9\x56\xaa\xa9\x56\xa8\x0f'
b'\x02\xaa\x55\xaa\xa5\x5a\xa8\x0f\x02\xaa\x55\x6a\x95\x5a\xa8\x0f'
b'\x02\xaa\x95\x5a\x95\x6a\xa8\x0f\x02\xaa\xa5\x56\x55\xaa\xa8\x0f'
b'\x02\xaa\xa9\x55\x56\xaa\xa8\x0f\x02\xaa\xaa\x55\x56\xaa\xa8\x0f'
b'\x02\xaa\xaa\x95\x5a\xaa\xa8\x0f\x02\xaa\xaa\xa5\x6a\xaa\xa8\x0f'
b'\x02\xaa\xaa\xa9\x6a\xaa\xa8\x0f\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f'
b'\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f'
b'\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f\x02\xaa\xaa\xaa\xaa\xaa\xa8\x0f'
b'\x00\x00\x00\x00\x00\x00\x00\x3f\xc0\x00\x00\x00\x00\x00\x00\xff'
),
1: (
b'\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff'
b'\xff\xff\xff\xff\xff\xff\xff\xff\xc0\x00\x00\x00\x00\x00\x00\xff'
b'\x00\x00\x00\x00\x00\x00\x00\x3f\x02\xaa\xaa\xaa\xaa\xaa\xa0\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x02\xaa\xaa\xaa\xaa\xaa\xa0\x3f'
b'\x00\x00\x00\x00\x00\x00\x00\x3f\xc0\x00\x00\x00\x00\x00\x00\xff'
),
2: (
b'\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff'
b'\xff\xff\xff\xff\xff\xff\xff\xff\xc0\x00\x00\x00\x00\x00\x00\xff'
b'\x00\x00\x00\x00\x00\x00\x00\x3f\x02\xaa\xaa\xaa\xaa\xaa\xa0\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xa5\x55\x56\xaa\xa8\x3f'
b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\x95\x55\x55\xaa\xa8\x3f'
b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\x95\x55\x55\xaa\xa8\x3f'
b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\x95\x55\x55\xaa\xa8\x3f'
b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\x95\x55\x55\xaa\xa8\x3f'
b'\x0a\xaa\x95\x55\x55\xaa\xa8\x3f\x0a\xaa\xa5\x55\x56\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xa8\x3f\x02\xaa\xaa\xaa\xaa\xaa\xa0\x3f'
b'\x00\x00\x00\x00\x00\x00\x00\x3f\xc0\x00\x00\x00\x00\x00\x00\xff'
),
3: (
b'\xff\xff\xff\x00\x00\xff\xff\xff\xff\xff\xc0\x00\x00\x03\xff\xff'
b'\xff\xfc\x00\xaa\xaa\x00\x3f\xff\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff'
b'\xff\xc0\xaa\xaa\xaa\xaa\x03\xff\xff\x02\xaa\xaa\xaa\xaa\x80\xff'
b'\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f'
b'\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f\xc0\xaa\xaa\xaa\xaa\xa6\xaa\x03'
b'\xc2\xaa\xaa\xaa\xaa\x95\xaa\x83\xc2\xaa\xaa\xaa\xaa\x56\xaa\x83'
b'\x0a\xaa\xaa\xaa\xa9\x56\xaa\x80\x0a\xaa\xaa\xaa\xa9\x5a\xaa\xa0'
b'\x0a\xaa\xaa\xaa\xa5\x6a\xaa\xa0\x0a\xaa\xaa\xaa\x95\x6a\xaa\xa0'
b'\x0a\xaa\xaa\xaa\x95\xaa\xaa\xa0\x0a\xaa\x96\xaa\x55\xaa\xaa\xa0'
b'\x0a\xaa\x95\xaa\x56\xaa\xaa\xa0\x0a\xaa\x95\x69\x5a\xaa\xaa\x80'
b'\xc2\xaa\xa5\x55\x5a\xaa\xaa\x83\xc2\xaa\xa9\x55\x6a\xaa\xaa\x83'
b'\xc0\xaa\xaa\x95\xaa\xaa\xaa\x03\xf0\xaa\xaa\xa5\xaa\xaa\xaa\x0f'
b'\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f'
b'\xff\x02\xaa\xaa\xaa\xaa\x80\xff\xff\xc0\xaa\xaa\xaa\xaa\x03\xff'
b'\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff\xff\xfc\x00\x2a\xa8\x00\x3f\xff'
b'\xff\xff\xc0\x00\x00\x03\xff\xff\xff\xff\xff\x00\x00\xff\xff\xff'
),
4: (
b'\xff\xff\xff\x00\x00\xff\xff\xff\xff\xff\xc0\x00\x00\x03\xff\xff'
b'\xff\xfc\x00\xaa\xaa\x00\x3f\xff\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff'
b'\xff\xc0\xaa\xaa\xaa\xaa\x03\xff\xff\x02\xaa\xaa\xaa\xaa\x80\xff'
b'\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f'
b'\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03'
b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0'
b'\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0'
b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83'
b'\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f'
b'\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f'
b'\xff\x02\xaa\xaa\xaa\xaa\x80\xff\xff\xc0\xaa\xaa\xaa\xaa\x03\xff'
b'\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff\xff\xfc\x00\xaa\xaa\x00\x3f\xff'
b'\xff\xff\xc0\x00\x00\x03\xff\xff\xff\xff\xff\x00\x00\xff\xff\xff'
),
5: (
b'\xff\xff\xff\x00\x00\xff\xff\xff\xff\xff\xc0\x00\x00\x03\xff\xff'
b'\xff\xfc\x00\xaa\xaa\x00\x3f\xff\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff'
b'\xff\xc0\xaa\xaa\xaa\xaa\x03\xff\xff\x02\xaa\xaa\xaa\xaa\x80\xff'
b'\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f'
b'\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03'
b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83'
b'\x0a\xaa\xaa\xa5\x6a\xaa\xaa\xa0\x0a\xaa\xaa\x95\x5a\xaa\xaa\xa0'
b'\x0a\xaa\xaa\x55\x56\xaa\xaa\xa0\x0a\xaa\xaa\x55\x56\xaa\xaa\xa0'
b'\x0a\xaa\xaa\x55\x56\xaa\xaa\xa0\x0a\xaa\xaa\x95\x5a\xaa\xaa\xa0'
b'\x0a\xaa\xaa\xa5\x6a\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0'
b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83'
b'\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f'
b'\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f'
b'\xff\x02\xaa\xaa\xaa\xaa\x80\xff\xff\xc0\xaa\xaa\xaa\xaa\x03\xff'
b'\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff\xff\xfc\x00\xaa\xaa\x00\x3f\xff'
b'\xff\xff\xc0\x00\x00\x03\xff\xff\xff\xff\xff\x00\x00\xff\xff\xff'
),
6: (
b'\xff\xff\xff\xfd\x7f\xff\xff\xff\xff\xff\xc0\xf5\x5f\x03\xff\xff'
b'\xff\xfc\x00\xf5\x5f\x00\x3f\xff\xff\xf0\x0a\xa5\x5a\xa0\x0f\xff'
b'\xff\xc0\xaa\xa5\x5a\xaa\x03\xff\xff\x02\xaa\xa5\x5a\xaa\x80\xff'
b'\xfc\x0a\xaa\xa5\x5a\xaa\xa0\x3f\xf0\x2a\xaa\xa5\x5a\xaa\xa8\x0f'
b'\xf0\xaa\xaa\xa5\x5a\xaa\xaa\x0f\xc0\xaa\xaa\xa5\x5a\xaa\xaa\x03'
b'\xc2\xaa\xaa\xa5\x5a\xaa\xaa\x83\xc2\xaa\xaa\xa5\x5a\xaa\xaa\x83'
b'\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0'
b'\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0'
b'\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0\x0a\xaa\xaa\xa5\x5a\xaa\xaa\xa0'
b'\x0a\xaa\xaa\xa9\x6a\xaa\xaa\xa0\x0a\xaa\xaa\xaa\xaa\xaa\xaa\xa0'
b'\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83\xc2\xaa\xaa\xaa\xaa\xaa\xaa\x83'
b'\xc0\xaa\xaa\xaa\xaa\xaa\xaa\x03\xf0\xaa\xaa\xaa\xaa\xaa\xaa\x0f'
b'\xf0\x2a\xaa\xaa\xaa\xaa\xa8\x0f\xfc\x0a\xaa\xaa\xaa\xaa\xa0\x3f'
b'\xff\x02\xaa\xaa\xaa\xaa\x80\xff\xff\xc0\xaa\xaa\xaa\xaa\x03\xff'
b'\xff\xf0\x0a\xaa\xaa\xa0\x0f\xff\xff\xfc\x00\xaa\xaa\x00\x3f\xff'
b'\xff\xff\xc0\x00\x00\x03\xff\xff\xff\xff\xff\x00\x00\xff\xff\xff'
),
}
colortable = {
0: (b'\x00\x00\x00\x02\x00\x00\x00\x00\xff\xff\xff\x00\xff\xff\xff\x00'),
1: (b'\x00\x00\x00\x02\x01\x24\xf7\x00\xf0\xf0\xf0\x00\xff\xff\xff\x00'),
2: (b'\x00\x00\x00\x02\x00\xff\x00\x00\xf0\xf0\xf0\x00\xff\xff\xff\x00'),
3: (b'\x00\x00\x00\x02\x00\xce\xff\x00\xf0\xf0\xf0\x00\xff\xff\xff\x00'),
4: (b'\xff\xff\xff\x02\xff\xff\xff\x00\x00\x00\x00\x00\x00\x00\x00\x00'),
5: (b'\xff\xff\xff\x02\x01\x24\xf7\x00\x10\x10\x10\x00\x00\x00\x00\x00'),
6: (b'\xff\xff\xff\x02\x00\xff\x00\x00\x10\x10\x10\x00\x00\x00\x00\x00'),
7: (b'\xff\xff\xff\x02\x00\xce\xff\x00\x10\x10\x10\x00\x00\x00\x00\x00'),
}
width = 32
height = 32
colors = 2
def get_icon(icon_index = 0, color_index = 0):
return width, height, addressof(_icons[icon_index]), colors, addressof(colortable[color_index])
def draw(x, y, icon_index, draw_fct, color_index = 0):
draw_fct(x - width//2, y - height // 2, width, height, addressof(_icons[icon_index]), colors, addressof(colortable[color_index]))
| 60.261438 | 133 | 0.67603 | 2,125 | 9,220 | 2.925647 | 0.036706 | 0.512466 | 0.502332 | 0.434293 | 0.924883 | 0.878398 | 0.809394 | 0.760817 | 0.729452 | 0.703394 | 0 | 0.148496 | 0.07679 | 9,220 | 152 | 134 | 60.657895 | 0.581884 | 0.003471 | 0 | 0.575342 | 1 | 0.821918 | 0.836146 | 0.836146 | 0 | 1 | 0 | 0 | 0 | 1 | 0.013699 | false | 0 | 0.006849 | 0.006849 | 0.027397 | 0 | 0 | 0 | 0 | null | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | null | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 14 |
7eba828afd143c2c50ba8c130be49d1be5cef779 | 3,737 | py | Python | tests/test_binder.py | erayon/cycli | 2947ad90396091652e3f43a9f61a74539e3f310e | [
"MIT"
] | 290 | 2015-07-31T07:07:11.000Z | 2021-12-15T09:39:59.000Z | tests/test_binder.py | erayon/cycli | 2947ad90396091652e3f43a9f61a74539e3f310e | [
"MIT"
] | 56 | 2015-08-01T14:47:15.000Z | 2019-05-14T11:43:03.000Z | tests/test_binder.py | erayon/cycli | 2947ad90396091652e3f43a9f61a74539e3f310e | [
"MIT"
] | 40 | 2015-08-07T13:30:40.000Z | 2021-05-02T00:58:09.000Z | from __future__ import unicode_literals
from mock import call, MagicMock
import pytest
from cycli.binder import *
@pytest.fixture
def binder():
return MagicMock()
def test_curly_left(binder):
curly_left(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call("{"), call("}", move_cursor=False)])
def test_curly_right_curly_current(binder):
binder.cli.current_buffer.document.current_char = "}"
curly_right(binder)
binder.cli.current_buffer.cursor_right.assert_called_once_with()
def test_curly_right_not_curly_current(binder):
binder.cli.current_buffer.document.current_char = "n"
curly_right(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call("}")])
def test_paren_left(binder):
paren_left(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call("("), call(")", move_cursor=False)])
def test_paren_right_paren_current(binder):
binder.cli.current_buffer.document.current_char = ")"
paren_right(binder)
binder.cli.current_buffer.cursor_right.assert_called_once_with()
def test_paren_right_not_paren_current(binder):
binder.cli.current_buffer.document.current_char = "n"
paren_right(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call(")")])
def test_bracket_left(binder):
bracket_left(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call("["), call("]", move_cursor=False)])
def test_bracket_right_bracket_current(binder):
binder.cli.current_buffer.document.current_char = "]"
bracket_right(binder)
binder.cli.current_buffer.cursor_right.assert_called_once_with()
def test_bracket_right_not_bracket_current(binder):
binder.cli.current_buffer.document.current_char = "n"
bracket_right(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call("]")])
def test_apostrophe_current(binder):
binder.cli.current_buffer.document.current_char = "'"
apostrophe(binder)
binder.cli.current_buffer.cursor_right.assert_called_once_with()
def test_apostrophe_not_current(binder):
binder.cli.current_buffer.document.current_char = "n"
apostrophe(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call("'"), call("'", move_cursor=False)])
def test_quote_current(binder):
binder.cli.current_buffer.document.current_char = "\""
quote(binder)
binder.cli.current_buffer.cursor_right.assert_called_once_with()
def test_quote_not_current(binder):
binder.cli.current_buffer.document.current_char = "n"
quote(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call("\""), call("\"", move_cursor=False)])
def test_backtick_current(binder):
binder.cli.current_buffer.document.current_char = "`"
backtick(binder)
binder.cli.current_buffer.cursor_right.assert_called_once_with()
def test_backtick_not_current(binder):
binder.cli.current_buffer.document.current_char = "n"
backtick(binder)
binder.cli.current_buffer.insert_text.assert_has_calls([call("`"), call("`", move_cursor=False)])
def test_backspace_no_pattern(binder):
binder.cli.current_buffer.document.current_char = "n"
backspace(binder)
binder.cli.current_buffer.delete_before_cursor.assert_called_once_with()
@pytest.mark.parametrize(("char_before_cursor", "current_char"), [
("(", ")"),
("[", "]"),
("{", "}"),
("'", "'"),
('"', '"'),
("`", "`")
])
def test_backspace_with_pattern(char_before_cursor, current_char):
binder = MagicMock()
binder.cli.current_buffer.document.char_before_cursor = char_before_cursor
binder.cli.current_buffer.document.current_char = current_char
backspace(binder)
binder.cli.current_buffer.cursor_right.assert_called_once_with()
binder.cli.current_buffer.delete_before_cursor.assert_called_once_with(2)
| 29.896 | 101 | 0.779235 | 509 | 3,737 | 5.333988 | 0.096267 | 0.109392 | 0.194475 | 0.267403 | 0.836832 | 0.805893 | 0.780111 | 0.752486 | 0.752486 | 0.675138 | 0 | 0.000293 | 0.086433 | 3,737 | 124 | 102 | 30.137097 | 0.794962 | 0 | 0 | 0.337349 | 0 | 0 | 0.020605 | 0 | 0 | 0 | 0 | 0 | 0.216867 | 1 | 0.216867 | false | 0 | 0.048193 | 0.012048 | 0.277108 | 0 | 0 | 0 | 0 | null | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 9 |
7ed019aa7c9507a8c3e573da8a623a8d1f824c03 | 6,651 | py | Python | Abt-Buy/rltk_exp/feature_vector.py | cybergla/rltk-experimentation | 30485068c557bd218f40cc7db1d05cb3e1e61642 | [
"MIT"
] | 1 | 2018-08-15T17:58:42.000Z | 2018-08-15T17:58:42.000Z | Abt-Buy/rltk_exp/feature_vector.py | cybergla/rltk-experimentation | 30485068c557bd218f40cc7db1d05cb3e1e61642 | [
"MIT"
] | null | null | null | Abt-Buy/rltk_exp/feature_vector.py | cybergla/rltk-experimentation | 30485068c557bd218f40cc7db1d05cb3e1e61642 | [
"MIT"
] | 5 | 2018-10-26T16:02:45.000Z | 2021-07-15T10:54:31.000Z | from create_datasets import *
# def generate_feature_vector(r_abt, r_buy):
# brand_score = 0.5
# if r_abt.brand_cleaned and r_buy.brand_cleaned:
# if r_abt.brand_cleaned == r_buy.brand_cleaned:
# brand_score = 1
# else:
# if len(r_abt.brand_cleaned) >= len(r_buy.brand_cleaned):
# common_str = r_buy.brand_cleaned
# if r_abt.brand_cleaned.startswith(common_str) or r_abt.brand_cleaned.endswith(common_str):
# brand_score = 1
# else:
# brand_score = 0
# model_score = 0.5
# if r_abt.model_cleaned and r_buy.model_cleaned:
# if r_abt.model_cleaned == r_buy.model_cleaned:
# model_score = 1
# else:
# if len(r_abt.model_cleaned) >= len(r_buy.model_cleaned):
# common_str = r_buy.model_cleaned
# if r_abt.model_cleaned.startswith(common_str) or r_abt.model_cleaned.endswith(common_str):
# model_score = 1
# else:
# model_score = 0
#
# if brand_score == 1 and model_score == 1:
# jaccard_score = 1
# else:
# jaccard_score = rltk.jaccard_index_similarity(r_abt.name_tokens, r_buy.name_tokens)
#
# return [brand_score, model_score, jaccard_score]
def generate_feature_vector_raw(r_abt, r_buy):
# brand
brand_score = None
if r_abt.brand_cleaned and r_buy.brand_cleaned:
if r_abt.brand_cleaned == r_buy.brand_cleaned:
brand_score = 1
# model 1
model_score = None
model_marker = 0
if r_abt.model_cleaned and r_buy.model_cleaned:
if r_abt.model_cleaned == r_buy.model_cleaned:
model_score = 1
else:
if len(r_abt.model_cleaned) > len(r_buy.model_cleaned):
if r_abt.model_cleaned.startswith(r_buy.model_cleaned) \
or r_abt.model_cleaned.endswith(r_buy.model_cleaned):
model_score = 1
else:
model_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.model_cleaned)
elif len(r_abt.model_cleaned) < len(r_buy.model_cleaned):
if r_buy.model_cleaned.startswith(r_abt.model_cleaned) \
or r_buy.model_cleaned.endswith(r_abt.model_cleaned):
model_score = 1
else:
model_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.model_cleaned)
else:
model_score = 0
# model 2
model2_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.name_suffix)
# name tokens jaccard
jaccard_score = rltk.jaccard_index_similarity(r_abt.name_tokens, r_buy.name_tokens)
# name tokens tf-idf
# t_x = collections.Counter(r_abt.name_tokens)
# tf_x = {k: float(v) / len(r_abt.name_tokens) for k, v in t_x.items()}
# tfidf_x = {k : tf_x[k] / idf[k] for k, v in tf_x.items()}
# t_y = collections.Counter(r_buy.name_tokens)
# tf_y = {k: float(v) / len(r_buy.name_tokens) for k, v in t_y.items()}
# tfidf_y = {k : tf_y[k] / idf[k] for k, v in tf_y.items()}
# tfidf_score = rltk.tf_idf_similarity_by_dict(tfidf_x, tfidf_y)
tfidf_score = tfidf.similarity(r_abt.id, r_buy.id)
# price
if r_abt.price and r_buy.price:
price_marker = 1
abt_price = float(r_abt.price)
buy_price = float(r_buy.price)
if abt_price == 0 and buy_price == 0:
price_difference = 0
else:
price_difference = float(abs(abt_price - buy_price)) / max(abt_price, buy_price)
else:
price_marker = 0
price_difference = 0
return [brand_score, model_score,
model2_score, jaccard_score, tfidf_score, price_difference, price_marker]
def generate_feature_vector(r_abt, r_buy):
# brand
brand_score = 0.2
brand_marker = 0
if r_abt.brand_cleaned and r_buy.brand_cleaned:
if r_abt.brand_cleaned == r_buy.brand_cleaned:
brand_score = 1
brand_marker = 1
# model 1
model_score = 0.2
model_marker = 0
if r_abt.model_cleaned and r_buy.model_cleaned:
if r_abt.model_cleaned == r_buy.model_cleaned:
model_score = 1
model_marker = 1
else:
if len(r_abt.model_cleaned) > len(r_buy.model_cleaned):
if r_abt.model_cleaned.startswith(r_buy.model_cleaned) \
or r_abt.model_cleaned.endswith(r_buy.model_cleaned):
model_score = 1
model_marker = 1
else:
model_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.model_cleaned)
elif len(r_abt.model_cleaned) < len(r_buy.model_cleaned):
if r_buy.model_cleaned.startswith(r_abt.model_cleaned) \
or r_buy.model_cleaned.endswith(r_abt.model_cleaned):
model_score = 1
model_marker = 1
else:
model_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.model_cleaned)
else:
model_score = 0
# model 2
model2_score = rltk.levenshtein_similarity(r_abt.model_cleaned, r_buy.name_suffix)
# name tokens jaccard
jaccard_score = rltk.jaccard_index_similarity(r_abt.name_tokens, r_buy.name_tokens)
tfidf_score = tfidf.similarity(r_abt.id, r_buy.id)
# price
if r_abt.price and r_buy.price:
price_marker = 1
abt_price = float(r_abt.price)
buy_price = float(r_buy.price)
if abt_price == 0 and buy_price == 0:
price_difference = 0
else:
price_difference = float(abs(abt_price - buy_price)) / max(abt_price, buy_price)
else:
price_marker = 0
price_difference = 0
return [brand_score, brand_marker, model_score, model_marker,
model2_score, jaccard_score, tfidf_score, price_difference, price_marker]
def non_ml_method(r_abt, r_buy):
brand_score = 0
if r_abt.brand_cleaned and r_buy.brand_cleaned:
if r_abt.brand_cleaned == r_buy.brand_cleaned:
brand_score = 1
model_score = 0
if r_abt.model_cleaned and r_buy.model_cleaned:
if r_abt.model_cleaned == r_buy.model_cleaned:
model_score = 1
jaccard_score = rltk.jaccard_index_similarity(r_abt.name_tokens, r_buy.name_tokens)
if model_score == 1:
return True
total = brand_score * 0.3 + model_score * 0.3 + jaccard_score * 0.4
return total > 0.45 | 38.445087 | 108 | 0.621711 | 948 | 6,651 | 3.996835 | 0.074895 | 0.059119 | 0.068884 | 0.12246 | 0.86408 | 0.825812 | 0.811824 | 0.773291 | 0.765901 | 0.73819 | 0 | 0.014487 | 0.294241 | 6,651 | 173 | 109 | 38.445087 | 0.792714 | 0.266576 | 0 | 0.833333 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.029412 | false | 0 | 0.009804 | 0 | 0.078431 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 |
bc37c9139bda28d7dcbe08d115fee4810065a313 | 14,971 | py | Python | environments/CTF/loggingFunctions.py | vanstrn/RL_public | 0e971e40e063b17918460e19728f95d7924af8db | [
"MIT"
] | 1 | 2021-03-19T17:57:51.000Z | 2021-03-19T17:57:51.000Z | environments/CTF/loggingFunctions.py | vanstrn/RL_public | 0e971e40e063b17918460e19728f95d7924af8db | [
"MIT"
] | null | null | null | environments/CTF/loggingFunctions.py | vanstrn/RL_public | 0e971e40e063b17918460e19728f95d7924af8db | [
"MIT"
] | null | null | null | import tensorflow as tf
import numpy as np
import itertools
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import random
class ValueTest_StackedStates(tf.keras.callbacks.Callback):
def __init__(self,superEpochs,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[3]
self.imageDir=imageDir
self.freq = freq
self.superEpochs = superEpochs
def on_train_end(self, logs=None):
if self.superEpochs%self.freq == 0:
s=self.env.reset()
rewardMap = np.zeros((s.shape[1],s.shape[2]))
for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])):
grid = self.env.ConstructSample([i,j])
if grid is None: continue
tmp = np.concatenate([grid] * 4, axis=2)
[value] = SF4.predict(np.expand_dims(tmp,0))
rewardMap[i,j] = value
fig=plt.figure(figsize=(5.5, 8))
fig.add_subplot(2,1,1)
plt.title("State")
imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1)
fig.add_subplot(2,1,2)
plt.title("Value Prediction")
imgplot = plt.imshow(rewardMap)
fig.colorbar(imgplot)
plt.savefig(self.imageDir+"/ValuePred"+str(self.superEpochs)+".png")
plt.close()
class ValueTest(tf.keras.callbacks.Callback):
def __init__(self,superEpochs,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[3]
self.imageDir=imageDir
self.freq = freq
self.superEpochs = superEpochs
def on_train_end(self, logs=None):
if self.superEpochs%self.freq == 0:
s=self.env.reset()
rewardMap = np.zeros((s.shape[1],s.shape[2]))
for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])):
grid = self.env.ConstructSample([i,j])
if grid is None: continue
[value] = self.network.predict(np.expand_dims(grid,0))
rewardMap[i,j] = value
fig=plt.figure(figsize=(5.5, 8))
fig.add_subplot(2,1,1)
plt.title("State")
imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1)
fig.add_subplot(2,1,2)
plt.title("Value Prediction")
imgplot = plt.imshow(rewardMap)
fig.colorbar(imgplot)
plt.savefig(self.imageDir+"/ValuePred"+str(self.superEpochs)+".png")
plt.close()
class ValueTest_enemies(tf.keras.callbacks.Callback):
def __init__(self,superEpochs,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[3]
self.imageDir=imageDir
self.freq = freq
self.superEpochs = superEpochs
def on_train_end(self, logs=None):
if self.superEpochs%self.freq == 0:
s = self.env.reset()
for sample in range(5):
enemy_position = [random.randint(0,s.shape[1]-1),random.randint(0,s.shape[2]-1)]
while s[0,enemy_position[0],enemy_position[1],3] ==1:
enemy_position = [random.randint(0,s.shape[1]-1),random.randint(0,s.shape[2]-1)]
rewardMap = np.zeros((s.shape[1],s.shape[2]))
for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])):
grid = self.env.ConstructSample_e([i,j],enemy_position)
if grid is None: continue
[value] = self.network.predict(np.expand_dims(grid,0))
rewardMap[i,j] = value
fig=plt.figure(figsize=(5.5, 10))
fig.add_subplot(3,1,1)
plt.title("State Flags")
imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1)
fig.add_subplot(3,1,2)
plt.title("State Enemies")
imgplot = plt.imshow(grid[:,:,4], vmin=-1, vmax=1)
fig.add_subplot(3,1,3)
plt.title("Value Prediction Epoch "+str(self.superEpochs))
imgplot = plt.imshow(rewardMap)
fig.colorbar(imgplot)
plt.savefig(self.imageDir+"/ValuePred"+str(self.superEpochs)+"_sample"+str(sample)+".png")
plt.close()
class ImageGenerator_StackedStates(tf.keras.callbacks.Callback):
def __init__(self,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[0]
self.imageDir = imageDir
self.freq = freq
def on_epoch_end(self,epoch, logs=None):
if epoch%self.freq == 0:
state = self.env.reset()
[state_new,reward] = self.network.predict([np.stack([[0,0,0,0,0]]),state])
fig=plt.figure(figsize=(16, 8))
fig.add_subplot(2,4,1)
plt.title("State Territory")
imgplot = plt.imshow(state[0,:,:,1], vmin=-2, vmax=2)
fig.add_subplot(2,4,2)
plt.title("State Flags")
imgplot = plt.imshow(state[0,:,:,2], vmin=-10, vmax=10)
fig.add_subplot(2,4,3)
plt.title("State Obstacles")
imgplot = plt.imshow(state[0,:,:,3], vmin=-2, vmax=2)
fig.add_subplot(2,4,4)
plt.title("State Agents")
imgplot = plt.imshow(state[0,:,:,4],vmin=-10, vmax=10)
fig.add_subplot(2,4,5)
plt.title("Predicted Next State Territory")
imgplot = plt.imshow(state_new[0,:,:,1],vmin=-2, vmax=2)
fig.add_subplot(2,4,6)
plt.title("Predicted Next State Flags")
imgplot = plt.imshow(state_new[0,:,:,2],vmin=-10, vmax=10)
fig.add_subplot(2,4,7)
plt.title("Predicted Next State Obstacles")
imgplot = plt.imshow(state_new[0,:,:,3],vmin=-2, vmax=2)
fig.add_subplot(2,4,8)
plt.title("Predicted Next State Agents")
imgplot = plt.imshow(state_new[0,:,:,4],vmin=-10, vmax=10)
plt.savefig(self.imageDir+"/StatePredEpoch"+str(epoch)+".png")
class ImageGenerator_actions(tf.keras.callbacks.Callback):
def __init__(self,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[0]
self.imageDir = imageDir
self.freq = freq
def on_epoch_end(self,epoch, logs=None):
if epoch%self.freq == 0:
state = self.env.reset()
[state_new,reward] = self.network.predict([np.stack([[0,0,0,0,0]]),state])
fig=plt.figure(figsize=(16, 8))
fig.add_subplot(2,4,1)
plt.title("State Territory")
imgplot = plt.imshow(state[0,:,:,1], vmin=-2, vmax=2)
fig.add_subplot(2,4,2)
plt.title("State Flags")
imgplot = plt.imshow(state[0,:,:,2], vmin=-10, vmax=10)
fig.add_subplot(2,4,3)
plt.title("State Obstacles")
imgplot = plt.imshow(state[0,:,:,3], vmin=-2, vmax=2)
fig.add_subplot(2,4,4)
plt.title("State Agents")
imgplot = plt.imshow(state[0,:,:,4],vmin=-10, vmax=10)
fig.add_subplot(2,4,5)
plt.title("Predicted Next State Territory")
imgplot = plt.imshow(state_new[0,:,:,1],vmin=-2, vmax=2)
fig.add_subplot(2,4,6)
plt.title("Predicted Next State Flags")
imgplot = plt.imshow(state_new[0,:,:,2],vmin=-10, vmax=10)
fig.add_subplot(2,4,7)
plt.title("Predicted Next State Obstacles")
imgplot = plt.imshow(state_new[0,:,:,3],vmin=-2, vmax=2)
fig.add_subplot(2,4,8)
plt.title("Predicted Next State Agents")
imgplot = plt.imshow(state_new[0,:,:,4],vmin=-10, vmax=10)
plt.savefig(self.imageDir+"/StatePredEpoch"+str(epoch)+".png")
plt.close()
class ImageGenerator(tf.keras.callbacks.Callback):
def __init__(self,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[0]
self.imageDir = imageDir
self.freq = freq
def on_epoch_end(self,epoch, logs=None):
if epoch%self.freq == 0:
state = self.env.reset()
[state_new,reward] = self.network.predict([state])
fig=plt.figure(figsize=(16, 8))
fig.add_subplot(2,4,1)
plt.title("State Territory")
imgplot = plt.imshow(state[0,:,:,1], vmin=-2, vmax=2)
fig.add_subplot(2,4,2)
plt.title("State Flags")
imgplot = plt.imshow(state[0,:,:,2], vmin=-10, vmax=10)
fig.add_subplot(2,4,3)
plt.title("State Obstacles")
imgplot = plt.imshow(state[0,:,:,3], vmin=-2, vmax=2)
fig.add_subplot(2,4,4)
plt.title("State Agents")
imgplot = plt.imshow(state[0,:,:,4],vmin=-10, vmax=10)
fig.add_subplot(2,4,5)
plt.title("Predicted Next State Territory")
imgplot = plt.imshow(state_new[0,:,:,1],vmin=-2, vmax=2)
fig.add_subplot(2,4,6)
plt.title("Predicted Next State Flags")
imgplot = plt.imshow(state_new[0,:,:,2],vmin=-10, vmax=10)
fig.add_subplot(2,4,7)
plt.title("Predicted Next State Obstacles")
imgplot = plt.imshow(state_new[0,:,:,3],vmin=-2, vmax=2)
fig.add_subplot(2,4,8)
plt.title("Predicted Next State Agents")
imgplot = plt.imshow(state_new[0,:,:,4],vmin=-10, vmax=10)
plt.savefig(self.imageDir+"/StatePredEpoch"+str(epoch)+".png")
plt.close()
class RewardTest(tf.keras.callbacks.Callback):
def __init__(self,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[0]
self.imageDir = imageDir
self.freq = freq
def on_epoch_end(self,epoch, logs=None):
if epoch%self.freq == 0:
s = self.env.reset()
rewardMap = np.zeros((s.shape[1],s.shape[2]))
for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])):
grid = self.env.ConstructSample([i,j])
if grid is None: continue
[state_new,reward] = self.network.predict(np.expand_dims(grid,0))
rewardMap[i,j] = reward
fig=plt.figure(figsize=(5.5, 8))
fig.add_subplot(2,1,1)
plt.title("State")
imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1)
fig.add_subplot(2,1,2)
plt.title("Reward Prediction Epoch "+str(epoch))
imgplot = plt.imshow(rewardMap)
fig.colorbar(imgplot)
plt.savefig(self.imageDir+"/RewardPred"+str(epoch)+".png")
plt.close()
class RewardTest_actions(tf.keras.callbacks.Callback):
def __init__(self,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[0]
self.imageDir = imageDir
self.freq = freq
def on_epoch_end(self,epoch, logs=None):
if epoch%self.freq == 0:
s = self.env.reset()
rewardMap = np.zeros((s.shape[1],s.shape[2]))
for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])):
grid = self.env.ConstructSample([i,j])
if grid is None: continue
[state_new,reward] = self.network.predict([np.stack([[0,0,0,0,0]]),np.expand_dims(grid,0)])
rewardMap[i,j] = reward
fig=plt.figure(figsize=(5.5, 8))
fig.add_subplot(2,1,1)
plt.title("State")
imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1)
fig.add_subplot(2,1,2)
plt.title("Reward Prediction Epoch "+str(epoch))
imgplot = plt.imshow(rewardMap)
fig.colorbar(imgplot)
plt.savefig(self.imageDir+"/RewardPred"+str(epoch)+".png")
plt.close()
class RewardTest_actions_enemies(tf.keras.callbacks.Callback):
def __init__(self,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[0]
self.imageDir = imageDir
self.freq = freq
def on_epoch_end(self,epoch, logs=None):
if epoch%self.freq == 0:
s = self.env.reset()
for sample in range(5):
enemy_position = [random.randint(0,s.shape[1]-1),random.randint(0,s.shape[2]-1)]
while s[0,enemy_position[0],enemy_position[1],3] ==1:
enemy_position = [random.randint(0,s.shape[1]-1),random.randint(0,s.shape[2]-1)]
rewardMap = np.zeros((s.shape[1],s.shape[2]))
for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])):
grid = self.env.ConstructSample_e([i,j],enemy_position)
if grid is None: continue
[state_new,reward] = self.network.predict([np.stack([[0,0,0,0,0]]),np.expand_dims(grid,0)])
rewardMap[i,j] = reward
# fig=plt.figure()
fig=plt.figure(figsize=(5.5, 10))
fig.add_subplot(3,1,1)
plt.title("State Flags")
imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1)
fig.add_subplot(3,1,2)
plt.title("State Enemies")
imgplot = plt.imshow(grid[:,:,4], vmin=-1, vmax=1)
fig.add_subplot(3,1,3)
plt.title("Reward Prediction Epoch "+str(epoch))
imgplot = plt.imshow(rewardMap)
fig.colorbar(imgplot)
plt.savefig(self.imageDir+"/RewardPred"+str(epoch)+"_sample"+str(sample)+".png")
plt.close()
class RewardTest_StackedStates(tf.keras.callbacks.Callback):
def __init__(self,env,network,imageDir=None,freq=50):
self.env = env
self.network=network[0]
self.imageDir = imageDir
self.freq = freq
def on_epoch_end(self,epoch, logs=None):
if epoch%self.freq == 0:
s = self.env.reset()
rewardMap = np.zeros((s.shape[1],s.shape[2]))
for i,j in itertools.product(range(s.shape[1]),range(s.shape[2])):
grid = self.env.ConstructSample([i,j])
if grid is None: continue
tmp = np.concatenate([grid] * 4, axis=2)
[state_new,reward] = self.network.predict([np.stack([[1,1,1,1,1]]),np.expand_dims(tmp,0)])
rewardMap[i,j] = reward
fig=plt.figure(figsize=(5.5, 8))
fig.add_subplot(2,1,1)
plt.title("State")
imgplot = plt.imshow(self.env.get_obs_blue[:,:,2], vmin=-1, vmax=1)
fig.add_subplot(2,1,2)
plt.title("Reward Prediction Epoch "+str(epoch))
imgplot = plt.imshow(rewardMap)
fig.colorbar(imgplot)
plt.savefig(self.imageDir+"/RewardPred"+str(epoch)+".png")
plt.close()
| 45.643293 | 111 | 0.558079 | 2,022 | 14,971 | 4.051929 | 0.054896 | 0.057366 | 0.063469 | 0.058098 | 0.974612 | 0.969974 | 0.969974 | 0.961186 | 0.949469 | 0.949469 | 0 | 0.041284 | 0.28809 | 14,971 | 327 | 112 | 45.782875 | 0.727435 | 0.001069 | 0 | 0.925806 | 0 | 0 | 0.059854 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.064516 | false | 0 | 0.019355 | 0 | 0.116129 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
70c93acd37b7b2c527bb8dbf47471beb4e21276e | 68 | py | Python | recipes/recipes_emscripten/boost-histogram/test_import_boost_histogram.py | emscripten-forge/recipes | 62cb3e146abc8945ac210f38e4e47c080698eae5 | [
"MIT"
] | 1 | 2022-03-10T16:50:56.000Z | 2022-03-10T16:50:56.000Z | recipes/recipes_emscripten/boost-histogram/test_import_boost_histogram.py | emscripten-forge/recipes | 62cb3e146abc8945ac210f38e4e47c080698eae5 | [
"MIT"
] | 9 | 2022-03-18T09:26:38.000Z | 2022-03-29T09:21:51.000Z | recipes/recipes_emscripten/boost-histogram/test_import_boost_histogram.py | emscripten-forge/recipes | 62cb3e146abc8945ac210f38e4e47c080698eae5 | [
"MIT"
] | null | null | null |
def test_import_boost_histogram():
import boost_histogram
| 13.6 | 34 | 0.75 | 8 | 68 | 5.875 | 0.625 | 0.468085 | 0.851064 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.205882 | 68 | 5 | 35 | 13.6 | 0.87037 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.5 | true | 0 | 1 | 0 | 1.5 | 0 | 1 | 0 | 0 | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 8 |
cb05f03e29fb61dbf7e515fc6ca4be9e2be0a4f9 | 2,753 | py | Python | gilbo_experiments_trace_diag/griewank-counterex-tensorflow.py | BachiLi/autodiff_comp | d2e33db4bd4ac7630b6ee6cfc56b3411cc37e14d | [
"MIT"
] | 2 | 2020-04-15T21:02:55.000Z | 2020-05-07T00:37:21.000Z | gilbo_experiments_trace_diag/griewank-counterex-tensorflow.py | BachiLi/autodiff_comp | d2e33db4bd4ac7630b6ee6cfc56b3411cc37e14d | [
"MIT"
] | null | null | null | gilbo_experiments_trace_diag/griewank-counterex-tensorflow.py | BachiLi/autodiff_comp | d2e33db4bd4ac7630b6ee6cfc56b3411cc37e14d | [
"MIT"
] | 1 | 2020-05-06T17:13:56.000Z | 2020-05-06T17:13:56.000Z |
import numpy as np
import tensorflow as tf
import time
import math
tf.compat.v1.enable_eager_execution()
def run_diag_trace(N,K):
def f(x):
A = tf.linalg.diag(x)
v = tf.linalg.trace(A)
for i in range(1,K):
v += tf.linalg.trace(A)
return v
def Df(x):
grads = None
with tf.GradientTape() as tape:
tape.watch(x)
v = f(x)
grads = tape.gradient(v, x)
return grads
x = tf.random.uniform([N])
f(x)
Df(x)
start_time = time.perf_counter()
v = f(x)
stop_time = time.perf_counter()
t = stop_time - start_time
start_time = time.perf_counter()
g = Df(x)
stop_time = time.perf_counter()
dt = stop_time - start_time
return (dt / t), t, dt
def xla_diag_trace(N,K):
@tf.function(
experimental_compile=True,
input_signature=[tf.TensorSpec(shape=[N], dtype=tf.float32)]
)
def f(x):
A = tf.linalg.diag(x)
v = tf.linalg.trace(A)
for i in range(1,K):
v += tf.linalg.trace(A)
return v
@tf.function(
experimental_compile=True,
input_signature=[tf.TensorSpec(shape=[N], dtype=tf.float32)]
)
def Df(x):
grads = None
with tf.GradientTape() as tape:
tape.watch(x)
A = tf.linalg.diag(x)
v = tf.linalg.trace(A)
for i in range(1,K):
v += tf.linalg.trace(A)
grads = tape.gradient(v, x)
return grads
x = tf.random.uniform([N])
f(x)
Df(x)
start_time = time.perf_counter()
v = f(x)
stop_time = time.perf_counter()
t = stop_time - start_time
start_time = time.perf_counter()
g = Df(x)
stop_time = time.perf_counter()
dt = stop_time - start_time
return (dt / t), t, dt
def xla_diag_dot(N,K):
@tf.function(
experimental_compile=True,
input_signature=[tf.TensorSpec(shape=[N], dtype=tf.float32)]
)
def f(x):
A = tf.linalg.diag(x)
v = tf.tensordot(A[:,0], A[0,:], 1)
return v
@tf.function(
experimental_compile=True,
input_signature=[tf.TensorSpec(shape=[N], dtype=tf.float32)]
)
def Df(x):
grads = None
with tf.GradientTape() as tape:
tape.watch(x)
A = tf.linalg.diag(x)
v = tf.tensordot(A[:,0], A[0,:], 1)
grads = tape.gradient(v, x)
return grads
x = tf.random.uniform([N])
f(x)
Df(x)
start_time = time.perf_counter()
v = f(x)
stop_time = time.perf_counter()
t = stop_time - start_time
start_time = time.perf_counter()
g = Df(x)
stop_time = time.perf_counter()
dt = stop_time - start_time
return (dt / t), t, dt
for K in range(1,11):
N = 4000 * K
ratio, t, dt = xla_diag_dot(N,K)
print(f"Griewank ratio for N = {N}: {ratio}, {t}, {dt}")
#print(f" : {t} {dt}")
print("\n----\n")
| 19.664286 | 64 | 0.587359 | 443 | 2,753 | 3.528217 | 0.148984 | 0.069098 | 0.092131 | 0.145873 | 0.880998 | 0.871401 | 0.871401 | 0.871401 | 0.871401 | 0.871401 | 0 | 0.012267 | 0.259717 | 2,753 | 139 | 65 | 19.805755 | 0.754661 | 0.007628 | 0 | 0.839623 | 0 | 0 | 0.020191 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.084906 | false | 0 | 0.037736 | 0 | 0.207547 | 0.018868 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
cb1332c53cf6e76166c0a29efa79493bd05d3f39 | 158 | py | Python | provision/src/util.py | SMISC/logging | b0c98737c46305e3e0718e429635043a4c57e9e0 | [
"MIT"
] | null | null | null | provision/src/util.py | SMISC/logging | b0c98737c46305e3e0718e429635043a4c57e9e0 | [
"MIT"
] | null | null | null | provision/src/util.py | SMISC/logging | b0c98737c46305e3e0718e429635043a4c57e9e0 | [
"MIT"
] | null | null | null | import datetime
def twittertime(dt):
return (datetime.datetime.strptime(dt, "%a %b %d %H:%M:%S +0000 %Y") - datetime.datetime(1970,1,1)).total_seconds()
| 31.6 | 119 | 0.683544 | 25 | 158 | 4.28 | 0.76 | 0.299065 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.071942 | 0.120253 | 158 | 4 | 120 | 39.5 | 0.697842 | 0 | 0 | 0 | 0 | 0 | 0.164557 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.333333 | false | 0 | 0.333333 | 0.333333 | 1 | 0 | 1 | 0 | 0 | null | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | 1 | 0 | 0 | 7 |
cb1a078cdd007b165ff0ae0c8e5db1b2f7e2b2ba | 105 | py | Python | token_myprint_runtest.py | jamesthiele/acodec | ac93cf5b6f7b1014e89abc47c0857e16d1a1aadf | [
"BSD-2-Clause"
] | null | null | null | token_myprint_runtest.py | jamesthiele/acodec | ac93cf5b6f7b1014e89abc47c0857e16d1a1aadf | [
"BSD-2-Clause"
] | null | null | null | token_myprint_runtest.py | jamesthiele/acodec | ac93cf5b6f7b1014e89abc47c0857e16d1a1aadf | [
"BSD-2-Clause"
] | null | null | null | import os
os.system("rm token_myprint_test.pyc")
import token_myprint_codec
import token_myprint_test
| 13.125 | 38 | 0.838095 | 17 | 105 | 4.823529 | 0.529412 | 0.439024 | 0.390244 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.104762 | 105 | 7 | 39 | 15 | 0.87234 | 0 | 0 | 0 | 0 | 0 | 0.240385 | 0.211538 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 0.75 | 0 | 0.75 | 0.75 | 1 | 0 | 0 | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 0 | 1 | 0 | 7 |
cb1ab9290d80e85d6655e191541d99167b4c4dff | 129,345 | py | Python | operators/keycloak-operator/python/pulumi_pulumi_kubernetes_crds_operators_keycloak_operator/keycloak/v1alpha1/outputs.py | pulumi/pulumi-kubernetes-crds | 372c4c0182f6b899af82d6edaad521aa14f22150 | [
"Apache-2.0"
] | null | null | null | operators/keycloak-operator/python/pulumi_pulumi_kubernetes_crds_operators_keycloak_operator/keycloak/v1alpha1/outputs.py | pulumi/pulumi-kubernetes-crds | 372c4c0182f6b899af82d6edaad521aa14f22150 | [
"Apache-2.0"
] | 2 | 2020-09-18T17:12:23.000Z | 2020-12-30T19:40:56.000Z | operators/keycloak-operator/python/pulumi_pulumi_kubernetes_crds_operators_keycloak_operator/keycloak/v1alpha1/outputs.py | pulumi/pulumi-kubernetes-crds | 372c4c0182f6b899af82d6edaad521aa14f22150 | [
"Apache-2.0"
] | null | null | null | # coding=utf-8
# *** WARNING: this file was generated by crd2pulumi. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union
from ... import _utilities, _tables
from . import outputs
__all__ = [
'KeycloakBackupSpec',
'KeycloakBackupSpecAws',
'KeycloakBackupSpecInstanceSelector',
'KeycloakBackupSpecInstanceSelectorMatchExpressions',
'KeycloakBackupStatus',
'KeycloakClientSpec',
'KeycloakClientSpecClient',
'KeycloakClientSpecClientProtocolMappers',
'KeycloakClientSpecRealmSelector',
'KeycloakClientSpecRealmSelectorMatchExpressions',
'KeycloakClientStatus',
'KeycloakRealmSpec',
'KeycloakRealmSpecInstanceSelector',
'KeycloakRealmSpecInstanceSelectorMatchExpressions',
'KeycloakRealmSpecRealm',
'KeycloakRealmSpecRealmClients',
'KeycloakRealmSpecRealmClientsProtocolMappers',
'KeycloakRealmSpecRealmIdentityProviders',
'KeycloakRealmSpecRealmOverrides',
'KeycloakRealmSpecRealmUsers',
'KeycloakRealmSpecRealmUsersCredentials',
'KeycloakRealmSpecRealmUsersFederatedIdentities',
'KeycloakRealmStatus',
'KeycloakSpec',
'KeycloakSpecExternalAccess',
'KeycloakSpecExternalDatabase',
'KeycloakSpecKeycloakDeploymentSpec',
'KeycloakSpecKeycloakDeploymentSpecResources',
'KeycloakSpecMigration',
'KeycloakSpecMigrationBackups',
'KeycloakSpecPodDisruptionBudget',
'KeycloakSpecPostgresDeploymentSpec',
'KeycloakSpecPostgresDeploymentSpecResources',
'KeycloakStatus',
'KeycloakUserSpec',
'KeycloakUserSpecRealmSelector',
'KeycloakUserSpecRealmSelectorMatchExpressions',
'KeycloakUserSpecUser',
'KeycloakUserSpecUserCredentials',
'KeycloakUserSpecUserFederatedIdentities',
'KeycloakUserStatus',
]
@pulumi.output_type
class KeycloakBackupSpec(dict):
"""
KeycloakBackupSpec defines the desired state of KeycloakBackup.
"""
def __init__(__self__, *,
aws: Optional['outputs.KeycloakBackupSpecAws'] = None,
instance_selector: Optional['outputs.KeycloakBackupSpecInstanceSelector'] = None,
restore: Optional[bool] = None,
storage_class_name: Optional[str] = None):
"""
KeycloakBackupSpec defines the desired state of KeycloakBackup.
:param 'KeycloakBackupSpecAwsArgs' aws: If provided, an automatic database backup will be created on AWS S3 instead of a local Persistent Volume. If this property is not provided - a local Persistent Volume backup will be chosen.
:param 'KeycloakBackupSpecInstanceSelectorArgs' instance_selector: Selector for looking up Keycloak Custom Resources.
:param bool restore: Controls automatic restore behavior. Currently not implemented.
In the future this will be used to trigger automatic restore for a given KeycloakBackup. Each backup will correspond to a single snapshot of the database (stored either in a Persistent Volume or AWS). If a user wants to restore it, all he/she needs to do is to change this flag to true. Potentially, it will be possible to restore a single backup multiple times.
:param str storage_class_name: Name of the StorageClass for Postgresql Backup Persistent Volume Claim
"""
if aws is not None:
pulumi.set(__self__, "aws", aws)
if instance_selector is not None:
pulumi.set(__self__, "instance_selector", instance_selector)
if restore is not None:
pulumi.set(__self__, "restore", restore)
if storage_class_name is not None:
pulumi.set(__self__, "storage_class_name", storage_class_name)
@property
@pulumi.getter
def aws(self) -> Optional['outputs.KeycloakBackupSpecAws']:
"""
If provided, an automatic database backup will be created on AWS S3 instead of a local Persistent Volume. If this property is not provided - a local Persistent Volume backup will be chosen.
"""
return pulumi.get(self, "aws")
@property
@pulumi.getter(name="instanceSelector")
def instance_selector(self) -> Optional['outputs.KeycloakBackupSpecInstanceSelector']:
"""
Selector for looking up Keycloak Custom Resources.
"""
return pulumi.get(self, "instance_selector")
@property
@pulumi.getter
def restore(self) -> Optional[bool]:
"""
Controls automatic restore behavior. Currently not implemented.
In the future this will be used to trigger automatic restore for a given KeycloakBackup. Each backup will correspond to a single snapshot of the database (stored either in a Persistent Volume or AWS). If a user wants to restore it, all he/she needs to do is to change this flag to true. Potentially, it will be possible to restore a single backup multiple times.
"""
return pulumi.get(self, "restore")
@property
@pulumi.getter(name="storageClassName")
def storage_class_name(self) -> Optional[str]:
"""
Name of the StorageClass for Postgresql Backup Persistent Volume Claim
"""
return pulumi.get(self, "storage_class_name")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakBackupSpecAws(dict):
"""
If provided, an automatic database backup will be created on AWS S3 instead of a local Persistent Volume. If this property is not provided - a local Persistent Volume backup will be chosen.
"""
def __init__(__self__, *,
credentials_secret_name: Optional[str] = None,
encryption_key_secret_name: Optional[str] = None,
schedule: Optional[str] = None):
"""
If provided, an automatic database backup will be created on AWS S3 instead of a local Persistent Volume. If this property is not provided - a local Persistent Volume backup will be chosen.
:param str credentials_secret_name: Provides a secret name used for connecting to AWS S3 Service. The secret needs to be in the following form:
apiVersion: v1 kind: Secret metadata: name: <Secret name> type: Opaque stringData: AWS_S3_BUCKET_NAME: <S3 Bucket Name> AWS_ACCESS_KEY_ID: <AWS Access Key ID> AWS_SECRET_ACCESS_KEY: <AWS Secret Key>
For more information, please refer to the Operator documentation.
:param str encryption_key_secret_name: If provided, the database backup will be encrypted. Provides a secret name used for encrypting database data. The secret needs to be in the following form:
apiVersion: v1 kind: Secret metadata: name: <Secret name> type: Opaque stringData: GPG_PUBLIC_KEY: <GPG Public Key> GPG_TRUST_MODEL: <GPG Trust Model> GPG_RECIPIENT: <GPG Recipient>
For more information, please refer to the Operator documentation.
:param str schedule: If specified, it will be used as a schedule for creating a CronJob.
"""
if credentials_secret_name is not None:
pulumi.set(__self__, "credentials_secret_name", credentials_secret_name)
if encryption_key_secret_name is not None:
pulumi.set(__self__, "encryption_key_secret_name", encryption_key_secret_name)
if schedule is not None:
pulumi.set(__self__, "schedule", schedule)
@property
@pulumi.getter(name="credentialsSecretName")
def credentials_secret_name(self) -> Optional[str]:
"""
Provides a secret name used for connecting to AWS S3 Service. The secret needs to be in the following form:
apiVersion: v1 kind: Secret metadata: name: <Secret name> type: Opaque stringData: AWS_S3_BUCKET_NAME: <S3 Bucket Name> AWS_ACCESS_KEY_ID: <AWS Access Key ID> AWS_SECRET_ACCESS_KEY: <AWS Secret Key>
For more information, please refer to the Operator documentation.
"""
return pulumi.get(self, "credentials_secret_name")
@property
@pulumi.getter(name="encryptionKeySecretName")
def encryption_key_secret_name(self) -> Optional[str]:
"""
If provided, the database backup will be encrypted. Provides a secret name used for encrypting database data. The secret needs to be in the following form:
apiVersion: v1 kind: Secret metadata: name: <Secret name> type: Opaque stringData: GPG_PUBLIC_KEY: <GPG Public Key> GPG_TRUST_MODEL: <GPG Trust Model> GPG_RECIPIENT: <GPG Recipient>
For more information, please refer to the Operator documentation.
"""
return pulumi.get(self, "encryption_key_secret_name")
@property
@pulumi.getter
def schedule(self) -> Optional[str]:
"""
If specified, it will be used as a schedule for creating a CronJob.
"""
return pulumi.get(self, "schedule")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakBackupSpecInstanceSelector(dict):
"""
Selector for looking up Keycloak Custom Resources.
"""
def __init__(__self__, *,
match_expressions: Optional[Sequence['outputs.KeycloakBackupSpecInstanceSelectorMatchExpressions']] = None,
match_labels: Optional[Mapping[str, str]] = None):
"""
Selector for looking up Keycloak Custom Resources.
:param Sequence['KeycloakBackupSpecInstanceSelectorMatchExpressionsArgs'] match_expressions: matchExpressions is a list of label selector requirements. The requirements are ANDed.
:param Mapping[str, str] match_labels: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.
"""
if match_expressions is not None:
pulumi.set(__self__, "match_expressions", match_expressions)
if match_labels is not None:
pulumi.set(__self__, "match_labels", match_labels)
@property
@pulumi.getter(name="matchExpressions")
def match_expressions(self) -> Optional[Sequence['outputs.KeycloakBackupSpecInstanceSelectorMatchExpressions']]:
"""
matchExpressions is a list of label selector requirements. The requirements are ANDed.
"""
return pulumi.get(self, "match_expressions")
@property
@pulumi.getter(name="matchLabels")
def match_labels(self) -> Optional[Mapping[str, str]]:
"""
matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.
"""
return pulumi.get(self, "match_labels")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakBackupSpecInstanceSelectorMatchExpressions(dict):
"""
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
"""
def __init__(__self__, *,
key: str,
operator: str,
values: Optional[Sequence[str]] = None):
"""
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
:param str key: key is the label key that the selector applies to.
:param str operator: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.
:param Sequence[str] values: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.
"""
pulumi.set(__self__, "key", key)
pulumi.set(__self__, "operator", operator)
if values is not None:
pulumi.set(__self__, "values", values)
@property
@pulumi.getter
def key(self) -> str:
"""
key is the label key that the selector applies to.
"""
return pulumi.get(self, "key")
@property
@pulumi.getter
def operator(self) -> str:
"""
operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.
"""
return pulumi.get(self, "operator")
@property
@pulumi.getter
def values(self) -> Optional[Sequence[str]]:
"""
values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.
"""
return pulumi.get(self, "values")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakBackupStatus(dict):
"""
KeycloakBackupStatus defines the observed state of KeycloakBackup.
"""
def __init__(__self__, *,
message: str,
phase: str,
ready: bool,
secondary_resources: Optional[Mapping[str, Sequence[str]]] = None):
"""
KeycloakBackupStatus defines the observed state of KeycloakBackup.
:param str message: Human-readable message indicating details about current operator phase or error.
:param str phase: Current phase of the operator.
:param bool ready: True if all resources are in a ready state and all work is done.
:param Mapping[str, Sequence[str]] secondary_resources: A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ]
"""
pulumi.set(__self__, "message", message)
pulumi.set(__self__, "phase", phase)
pulumi.set(__self__, "ready", ready)
if secondary_resources is not None:
pulumi.set(__self__, "secondary_resources", secondary_resources)
@property
@pulumi.getter
def message(self) -> str:
"""
Human-readable message indicating details about current operator phase or error.
"""
return pulumi.get(self, "message")
@property
@pulumi.getter
def phase(self) -> str:
"""
Current phase of the operator.
"""
return pulumi.get(self, "phase")
@property
@pulumi.getter
def ready(self) -> bool:
"""
True if all resources are in a ready state and all work is done.
"""
return pulumi.get(self, "ready")
@property
@pulumi.getter(name="secondaryResources")
def secondary_resources(self) -> Optional[Mapping[str, Sequence[str]]]:
"""
A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ]
"""
return pulumi.get(self, "secondary_resources")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakClientSpec(dict):
"""
KeycloakClientSpec defines the desired state of KeycloakClient.
"""
def __init__(__self__, *,
client: 'outputs.KeycloakClientSpecClient',
realm_selector: 'outputs.KeycloakClientSpecRealmSelector'):
"""
KeycloakClientSpec defines the desired state of KeycloakClient.
:param 'KeycloakClientSpecClientArgs' client: Keycloak Client REST object.
:param 'KeycloakClientSpecRealmSelectorArgs' realm_selector: Selector for looking up KeycloakRealm Custom Resources.
"""
pulumi.set(__self__, "client", client)
pulumi.set(__self__, "realm_selector", realm_selector)
@property
@pulumi.getter
def client(self) -> 'outputs.KeycloakClientSpecClient':
"""
Keycloak Client REST object.
"""
return pulumi.get(self, "client")
@property
@pulumi.getter(name="realmSelector")
def realm_selector(self) -> 'outputs.KeycloakClientSpecRealmSelector':
"""
Selector for looking up KeycloakRealm Custom Resources.
"""
return pulumi.get(self, "realm_selector")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakClientSpecClient(dict):
"""
Keycloak Client REST object.
"""
def __init__(__self__, *,
client_id: str,
access: Optional[Mapping[str, bool]] = None,
admin_url: Optional[str] = None,
attributes: Optional[Mapping[str, str]] = None,
base_url: Optional[str] = None,
bearer_only: Optional[bool] = None,
client_authenticator_type: Optional[str] = None,
consent_required: Optional[bool] = None,
default_roles: Optional[Sequence[str]] = None,
description: Optional[str] = None,
direct_access_grants_enabled: Optional[bool] = None,
enabled: Optional[bool] = None,
frontchannel_logout: Optional[bool] = None,
full_scope_allowed: Optional[bool] = None,
id: Optional[str] = None,
implicit_flow_enabled: Optional[bool] = None,
name: Optional[str] = None,
node_re_registration_timeout: Optional[int] = None,
not_before: Optional[int] = None,
protocol: Optional[str] = None,
protocol_mappers: Optional[Sequence['outputs.KeycloakClientSpecClientProtocolMappers']] = None,
public_client: Optional[bool] = None,
redirect_uris: Optional[Sequence[str]] = None,
root_url: Optional[str] = None,
secret: Optional[str] = None,
service_accounts_enabled: Optional[bool] = None,
standard_flow_enabled: Optional[bool] = None,
surrogate_auth_required: Optional[bool] = None,
use_template_config: Optional[bool] = None,
use_template_mappers: Optional[bool] = None,
use_template_scope: Optional[bool] = None,
web_origins: Optional[Sequence[str]] = None):
"""
Keycloak Client REST object.
:param str client_id: Client ID.
:param Mapping[str, bool] access: Access options.
:param str admin_url: Application Admin URL.
:param Mapping[str, str] attributes: Client Attributes.
:param str base_url: Application base URL.
:param bool bearer_only: True if a client supports only Bearer Tokens.
:param str client_authenticator_type: What Client authentication type to use.
:param bool consent_required: True if Consent Screen is required.
:param Sequence[str] default_roles: Default Client roles.
:param str description: Client description.
:param bool direct_access_grants_enabled: True if Direct Grant is enabled.
:param bool enabled: Client enabled flag.
:param bool frontchannel_logout: True if this client supports Front Channel logout.
:param bool full_scope_allowed: True if Full Scope is allowed.
:param str id: Client ID. If not specified, automatically generated.
:param bool implicit_flow_enabled: True if Implicit flow is enabled.
:param str name: Client name.
:param int node_re_registration_timeout: Node registration timeout.
:param int not_before: Not Before setting.
:param str protocol: Protocol used for this Client.
:param Sequence['KeycloakClientSpecClientProtocolMappersArgs'] protocol_mappers: Protocol Mappers.
:param bool public_client: True if this is a public Client.
:param Sequence[str] redirect_uris: A list of valid Redirection URLs.
:param str root_url: Application root URL.
:param str secret: Client Secret. The Operator will automatically create a Secret based on this value.
:param bool service_accounts_enabled: True if Service Accounts are enabled.
:param bool standard_flow_enabled: True if Standard flow is enabled.
:param bool surrogate_auth_required: Surrogate Authentication Required option.
:param bool use_template_config: True to use a Template Config.
:param bool use_template_mappers: True to use Template Mappers.
:param bool use_template_scope: True to use Template Scope.
:param Sequence[str] web_origins: A list of valid Web Origins.
"""
pulumi.set(__self__, "client_id", client_id)
if access is not None:
pulumi.set(__self__, "access", access)
if admin_url is not None:
pulumi.set(__self__, "admin_url", admin_url)
if attributes is not None:
pulumi.set(__self__, "attributes", attributes)
if base_url is not None:
pulumi.set(__self__, "base_url", base_url)
if bearer_only is not None:
pulumi.set(__self__, "bearer_only", bearer_only)
if client_authenticator_type is not None:
pulumi.set(__self__, "client_authenticator_type", client_authenticator_type)
if consent_required is not None:
pulumi.set(__self__, "consent_required", consent_required)
if default_roles is not None:
pulumi.set(__self__, "default_roles", default_roles)
if description is not None:
pulumi.set(__self__, "description", description)
if direct_access_grants_enabled is not None:
pulumi.set(__self__, "direct_access_grants_enabled", direct_access_grants_enabled)
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
if frontchannel_logout is not None:
pulumi.set(__self__, "frontchannel_logout", frontchannel_logout)
if full_scope_allowed is not None:
pulumi.set(__self__, "full_scope_allowed", full_scope_allowed)
if id is not None:
pulumi.set(__self__, "id", id)
if implicit_flow_enabled is not None:
pulumi.set(__self__, "implicit_flow_enabled", implicit_flow_enabled)
if name is not None:
pulumi.set(__self__, "name", name)
if node_re_registration_timeout is not None:
pulumi.set(__self__, "node_re_registration_timeout", node_re_registration_timeout)
if not_before is not None:
pulumi.set(__self__, "not_before", not_before)
if protocol is not None:
pulumi.set(__self__, "protocol", protocol)
if protocol_mappers is not None:
pulumi.set(__self__, "protocol_mappers", protocol_mappers)
if public_client is not None:
pulumi.set(__self__, "public_client", public_client)
if redirect_uris is not None:
pulumi.set(__self__, "redirect_uris", redirect_uris)
if root_url is not None:
pulumi.set(__self__, "root_url", root_url)
if secret is not None:
pulumi.set(__self__, "secret", secret)
if service_accounts_enabled is not None:
pulumi.set(__self__, "service_accounts_enabled", service_accounts_enabled)
if standard_flow_enabled is not None:
pulumi.set(__self__, "standard_flow_enabled", standard_flow_enabled)
if surrogate_auth_required is not None:
pulumi.set(__self__, "surrogate_auth_required", surrogate_auth_required)
if use_template_config is not None:
pulumi.set(__self__, "use_template_config", use_template_config)
if use_template_mappers is not None:
pulumi.set(__self__, "use_template_mappers", use_template_mappers)
if use_template_scope is not None:
pulumi.set(__self__, "use_template_scope", use_template_scope)
if web_origins is not None:
pulumi.set(__self__, "web_origins", web_origins)
@property
@pulumi.getter(name="clientId")
def client_id(self) -> str:
"""
Client ID.
"""
return pulumi.get(self, "client_id")
@property
@pulumi.getter
def access(self) -> Optional[Mapping[str, bool]]:
"""
Access options.
"""
return pulumi.get(self, "access")
@property
@pulumi.getter(name="adminUrl")
def admin_url(self) -> Optional[str]:
"""
Application Admin URL.
"""
return pulumi.get(self, "admin_url")
@property
@pulumi.getter
def attributes(self) -> Optional[Mapping[str, str]]:
"""
Client Attributes.
"""
return pulumi.get(self, "attributes")
@property
@pulumi.getter(name="baseUrl")
def base_url(self) -> Optional[str]:
"""
Application base URL.
"""
return pulumi.get(self, "base_url")
@property
@pulumi.getter(name="bearerOnly")
def bearer_only(self) -> Optional[bool]:
"""
True if a client supports only Bearer Tokens.
"""
return pulumi.get(self, "bearer_only")
@property
@pulumi.getter(name="clientAuthenticatorType")
def client_authenticator_type(self) -> Optional[str]:
"""
What Client authentication type to use.
"""
return pulumi.get(self, "client_authenticator_type")
@property
@pulumi.getter(name="consentRequired")
def consent_required(self) -> Optional[bool]:
"""
True if Consent Screen is required.
"""
return pulumi.get(self, "consent_required")
@property
@pulumi.getter(name="defaultRoles")
def default_roles(self) -> Optional[Sequence[str]]:
"""
Default Client roles.
"""
return pulumi.get(self, "default_roles")
@property
@pulumi.getter
def description(self) -> Optional[str]:
"""
Client description.
"""
return pulumi.get(self, "description")
@property
@pulumi.getter(name="directAccessGrantsEnabled")
def direct_access_grants_enabled(self) -> Optional[bool]:
"""
True if Direct Grant is enabled.
"""
return pulumi.get(self, "direct_access_grants_enabled")
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
Client enabled flag.
"""
return pulumi.get(self, "enabled")
@property
@pulumi.getter(name="frontchannelLogout")
def frontchannel_logout(self) -> Optional[bool]:
"""
True if this client supports Front Channel logout.
"""
return pulumi.get(self, "frontchannel_logout")
@property
@pulumi.getter(name="fullScopeAllowed")
def full_scope_allowed(self) -> Optional[bool]:
"""
True if Full Scope is allowed.
"""
return pulumi.get(self, "full_scope_allowed")
@property
@pulumi.getter
def id(self) -> Optional[str]:
"""
Client ID. If not specified, automatically generated.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter(name="implicitFlowEnabled")
def implicit_flow_enabled(self) -> Optional[bool]:
"""
True if Implicit flow is enabled.
"""
return pulumi.get(self, "implicit_flow_enabled")
@property
@pulumi.getter
def name(self) -> Optional[str]:
"""
Client name.
"""
return pulumi.get(self, "name")
@property
@pulumi.getter(name="nodeReRegistrationTimeout")
def node_re_registration_timeout(self) -> Optional[int]:
"""
Node registration timeout.
"""
return pulumi.get(self, "node_re_registration_timeout")
@property
@pulumi.getter(name="notBefore")
def not_before(self) -> Optional[int]:
"""
Not Before setting.
"""
return pulumi.get(self, "not_before")
@property
@pulumi.getter
def protocol(self) -> Optional[str]:
"""
Protocol used for this Client.
"""
return pulumi.get(self, "protocol")
@property
@pulumi.getter(name="protocolMappers")
def protocol_mappers(self) -> Optional[Sequence['outputs.KeycloakClientSpecClientProtocolMappers']]:
"""
Protocol Mappers.
"""
return pulumi.get(self, "protocol_mappers")
@property
@pulumi.getter(name="publicClient")
def public_client(self) -> Optional[bool]:
"""
True if this is a public Client.
"""
return pulumi.get(self, "public_client")
@property
@pulumi.getter(name="redirectUris")
def redirect_uris(self) -> Optional[Sequence[str]]:
"""
A list of valid Redirection URLs.
"""
return pulumi.get(self, "redirect_uris")
@property
@pulumi.getter(name="rootUrl")
def root_url(self) -> Optional[str]:
"""
Application root URL.
"""
return pulumi.get(self, "root_url")
@property
@pulumi.getter
def secret(self) -> Optional[str]:
"""
Client Secret. The Operator will automatically create a Secret based on this value.
"""
return pulumi.get(self, "secret")
@property
@pulumi.getter(name="serviceAccountsEnabled")
def service_accounts_enabled(self) -> Optional[bool]:
"""
True if Service Accounts are enabled.
"""
return pulumi.get(self, "service_accounts_enabled")
@property
@pulumi.getter(name="standardFlowEnabled")
def standard_flow_enabled(self) -> Optional[bool]:
"""
True if Standard flow is enabled.
"""
return pulumi.get(self, "standard_flow_enabled")
@property
@pulumi.getter(name="surrogateAuthRequired")
def surrogate_auth_required(self) -> Optional[bool]:
"""
Surrogate Authentication Required option.
"""
return pulumi.get(self, "surrogate_auth_required")
@property
@pulumi.getter(name="useTemplateConfig")
def use_template_config(self) -> Optional[bool]:
"""
True to use a Template Config.
"""
return pulumi.get(self, "use_template_config")
@property
@pulumi.getter(name="useTemplateMappers")
def use_template_mappers(self) -> Optional[bool]:
"""
True to use Template Mappers.
"""
return pulumi.get(self, "use_template_mappers")
@property
@pulumi.getter(name="useTemplateScope")
def use_template_scope(self) -> Optional[bool]:
"""
True to use Template Scope.
"""
return pulumi.get(self, "use_template_scope")
@property
@pulumi.getter(name="webOrigins")
def web_origins(self) -> Optional[Sequence[str]]:
"""
A list of valid Web Origins.
"""
return pulumi.get(self, "web_origins")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakClientSpecClientProtocolMappers(dict):
def __init__(__self__, *,
config: Optional[Mapping[str, str]] = None,
consent_required: Optional[bool] = None,
consent_text: Optional[str] = None,
id: Optional[str] = None,
name: Optional[str] = None,
protocol: Optional[str] = None,
protocol_mapper: Optional[str] = None):
"""
:param Mapping[str, str] config: Config options.
:param bool consent_required: True if Consent Screen is required.
:param str consent_text: Text to use for displaying Consent Screen.
:param str id: Protocol Mapper ID.
:param str name: Protocol Mapper Name.
:param str protocol: Protocol to use.
:param str protocol_mapper: Protocol Mapper to use
"""
if config is not None:
pulumi.set(__self__, "config", config)
if consent_required is not None:
pulumi.set(__self__, "consent_required", consent_required)
if consent_text is not None:
pulumi.set(__self__, "consent_text", consent_text)
if id is not None:
pulumi.set(__self__, "id", id)
if name is not None:
pulumi.set(__self__, "name", name)
if protocol is not None:
pulumi.set(__self__, "protocol", protocol)
if protocol_mapper is not None:
pulumi.set(__self__, "protocol_mapper", protocol_mapper)
@property
@pulumi.getter
def config(self) -> Optional[Mapping[str, str]]:
"""
Config options.
"""
return pulumi.get(self, "config")
@property
@pulumi.getter(name="consentRequired")
def consent_required(self) -> Optional[bool]:
"""
True if Consent Screen is required.
"""
return pulumi.get(self, "consent_required")
@property
@pulumi.getter(name="consentText")
def consent_text(self) -> Optional[str]:
"""
Text to use for displaying Consent Screen.
"""
return pulumi.get(self, "consent_text")
@property
@pulumi.getter
def id(self) -> Optional[str]:
"""
Protocol Mapper ID.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter
def name(self) -> Optional[str]:
"""
Protocol Mapper Name.
"""
return pulumi.get(self, "name")
@property
@pulumi.getter
def protocol(self) -> Optional[str]:
"""
Protocol to use.
"""
return pulumi.get(self, "protocol")
@property
@pulumi.getter(name="protocolMapper")
def protocol_mapper(self) -> Optional[str]:
"""
Protocol Mapper to use
"""
return pulumi.get(self, "protocol_mapper")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakClientSpecRealmSelector(dict):
"""
Selector for looking up KeycloakRealm Custom Resources.
"""
def __init__(__self__, *,
match_expressions: Optional[Sequence['outputs.KeycloakClientSpecRealmSelectorMatchExpressions']] = None,
match_labels: Optional[Mapping[str, str]] = None):
"""
Selector for looking up KeycloakRealm Custom Resources.
:param Sequence['KeycloakClientSpecRealmSelectorMatchExpressionsArgs'] match_expressions: matchExpressions is a list of label selector requirements. The requirements are ANDed.
:param Mapping[str, str] match_labels: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.
"""
if match_expressions is not None:
pulumi.set(__self__, "match_expressions", match_expressions)
if match_labels is not None:
pulumi.set(__self__, "match_labels", match_labels)
@property
@pulumi.getter(name="matchExpressions")
def match_expressions(self) -> Optional[Sequence['outputs.KeycloakClientSpecRealmSelectorMatchExpressions']]:
"""
matchExpressions is a list of label selector requirements. The requirements are ANDed.
"""
return pulumi.get(self, "match_expressions")
@property
@pulumi.getter(name="matchLabels")
def match_labels(self) -> Optional[Mapping[str, str]]:
"""
matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.
"""
return pulumi.get(self, "match_labels")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakClientSpecRealmSelectorMatchExpressions(dict):
"""
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
"""
def __init__(__self__, *,
key: str,
operator: str,
values: Optional[Sequence[str]] = None):
"""
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
:param str key: key is the label key that the selector applies to.
:param str operator: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.
:param Sequence[str] values: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.
"""
pulumi.set(__self__, "key", key)
pulumi.set(__self__, "operator", operator)
if values is not None:
pulumi.set(__self__, "values", values)
@property
@pulumi.getter
def key(self) -> str:
"""
key is the label key that the selector applies to.
"""
return pulumi.get(self, "key")
@property
@pulumi.getter
def operator(self) -> str:
"""
operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.
"""
return pulumi.get(self, "operator")
@property
@pulumi.getter
def values(self) -> Optional[Sequence[str]]:
"""
values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.
"""
return pulumi.get(self, "values")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakClientStatus(dict):
"""
KeycloakClientStatus defines the observed state of KeycloakClient
"""
def __init__(__self__, *,
message: str,
phase: str,
ready: bool,
secondary_resources: Optional[Mapping[str, Sequence[str]]] = None):
"""
KeycloakClientStatus defines the observed state of KeycloakClient
:param str message: Human-readable message indicating details about current operator phase or error.
:param str phase: Current phase of the operator.
:param bool ready: True if all resources are in a ready state and all work is done.
:param Mapping[str, Sequence[str]] secondary_resources: A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ]
"""
pulumi.set(__self__, "message", message)
pulumi.set(__self__, "phase", phase)
pulumi.set(__self__, "ready", ready)
if secondary_resources is not None:
pulumi.set(__self__, "secondary_resources", secondary_resources)
@property
@pulumi.getter
def message(self) -> str:
"""
Human-readable message indicating details about current operator phase or error.
"""
return pulumi.get(self, "message")
@property
@pulumi.getter
def phase(self) -> str:
"""
Current phase of the operator.
"""
return pulumi.get(self, "phase")
@property
@pulumi.getter
def ready(self) -> bool:
"""
True if all resources are in a ready state and all work is done.
"""
return pulumi.get(self, "ready")
@property
@pulumi.getter(name="secondaryResources")
def secondary_resources(self) -> Optional[Mapping[str, Sequence[str]]]:
"""
A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ]
"""
return pulumi.get(self, "secondary_resources")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpec(dict):
"""
KeycloakRealmSpec defines the desired state of KeycloakRealm.
"""
def __init__(__self__, *,
realm: 'outputs.KeycloakRealmSpecRealm',
instance_selector: Optional['outputs.KeycloakRealmSpecInstanceSelector'] = None,
realm_overrides: Optional[Sequence['outputs.KeycloakRealmSpecRealmOverrides']] = None):
"""
KeycloakRealmSpec defines the desired state of KeycloakRealm.
:param 'KeycloakRealmSpecRealmArgs' realm: Keycloak Realm REST object.
:param 'KeycloakRealmSpecInstanceSelectorArgs' instance_selector: Selector for looking up Keycloak Custom Resources.
:param Sequence['KeycloakRealmSpecRealmOverridesArgs'] realm_overrides: A list of overrides to the default Realm behavior.
"""
pulumi.set(__self__, "realm", realm)
if instance_selector is not None:
pulumi.set(__self__, "instance_selector", instance_selector)
if realm_overrides is not None:
pulumi.set(__self__, "realm_overrides", realm_overrides)
@property
@pulumi.getter
def realm(self) -> 'outputs.KeycloakRealmSpecRealm':
"""
Keycloak Realm REST object.
"""
return pulumi.get(self, "realm")
@property
@pulumi.getter(name="instanceSelector")
def instance_selector(self) -> Optional['outputs.KeycloakRealmSpecInstanceSelector']:
"""
Selector for looking up Keycloak Custom Resources.
"""
return pulumi.get(self, "instance_selector")
@property
@pulumi.getter(name="realmOverrides")
def realm_overrides(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmOverrides']]:
"""
A list of overrides to the default Realm behavior.
"""
return pulumi.get(self, "realm_overrides")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecInstanceSelector(dict):
"""
Selector for looking up Keycloak Custom Resources.
"""
def __init__(__self__, *,
match_expressions: Optional[Sequence['outputs.KeycloakRealmSpecInstanceSelectorMatchExpressions']] = None,
match_labels: Optional[Mapping[str, str]] = None):
"""
Selector for looking up Keycloak Custom Resources.
:param Sequence['KeycloakRealmSpecInstanceSelectorMatchExpressionsArgs'] match_expressions: matchExpressions is a list of label selector requirements. The requirements are ANDed.
:param Mapping[str, str] match_labels: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.
"""
if match_expressions is not None:
pulumi.set(__self__, "match_expressions", match_expressions)
if match_labels is not None:
pulumi.set(__self__, "match_labels", match_labels)
@property
@pulumi.getter(name="matchExpressions")
def match_expressions(self) -> Optional[Sequence['outputs.KeycloakRealmSpecInstanceSelectorMatchExpressions']]:
"""
matchExpressions is a list of label selector requirements. The requirements are ANDed.
"""
return pulumi.get(self, "match_expressions")
@property
@pulumi.getter(name="matchLabels")
def match_labels(self) -> Optional[Mapping[str, str]]:
"""
matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.
"""
return pulumi.get(self, "match_labels")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecInstanceSelectorMatchExpressions(dict):
"""
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
"""
def __init__(__self__, *,
key: str,
operator: str,
values: Optional[Sequence[str]] = None):
"""
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
:param str key: key is the label key that the selector applies to.
:param str operator: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.
:param Sequence[str] values: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.
"""
pulumi.set(__self__, "key", key)
pulumi.set(__self__, "operator", operator)
if values is not None:
pulumi.set(__self__, "values", values)
@property
@pulumi.getter
def key(self) -> str:
"""
key is the label key that the selector applies to.
"""
return pulumi.get(self, "key")
@property
@pulumi.getter
def operator(self) -> str:
"""
operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.
"""
return pulumi.get(self, "operator")
@property
@pulumi.getter
def values(self) -> Optional[Sequence[str]]:
"""
values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.
"""
return pulumi.get(self, "values")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecRealm(dict):
"""
Keycloak Realm REST object.
"""
def __init__(__self__, *,
realm: str,
admin_events_details_enabled: Optional[bool] = None,
admin_events_enabled: Optional[bool] = None,
clients: Optional[Sequence['outputs.KeycloakRealmSpecRealmClients']] = None,
display_name: Optional[str] = None,
enabled: Optional[bool] = None,
events_enabled: Optional[bool] = None,
events_listeners: Optional[Sequence[str]] = None,
id: Optional[str] = None,
identity_providers: Optional[Sequence['outputs.KeycloakRealmSpecRealmIdentityProviders']] = None,
users: Optional[Sequence['outputs.KeycloakRealmSpecRealmUsers']] = None):
"""
Keycloak Realm REST object.
:param str realm: Realm name.
:param bool admin_events_details_enabled: Enable admin events details TODO: change to values and use kubebuilder default annotation once supported
:param bool admin_events_enabled: Enable events recording TODO: change to values and use kubebuilder default annotation once supported
:param Sequence['KeycloakRealmSpecRealmClientsArgs'] clients: A set of Keycloak Clients.
:param str display_name: Realm display name.
:param bool enabled: Realm enabled flag.
:param bool events_enabled: Enable events recording TODO: change to values and use kubebuilder default annotation once supported
:param Sequence[str] events_listeners: A set of Event Listeners.
:param Sequence['KeycloakRealmSpecRealmIdentityProvidersArgs'] identity_providers: A set of Identity Providers.
:param Sequence['KeycloakRealmSpecRealmUsersArgs'] users: A set of Keycloak Users.
"""
pulumi.set(__self__, "realm", realm)
if admin_events_details_enabled is not None:
pulumi.set(__self__, "admin_events_details_enabled", admin_events_details_enabled)
if admin_events_enabled is not None:
pulumi.set(__self__, "admin_events_enabled", admin_events_enabled)
if clients is not None:
pulumi.set(__self__, "clients", clients)
if display_name is not None:
pulumi.set(__self__, "display_name", display_name)
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
if events_enabled is not None:
pulumi.set(__self__, "events_enabled", events_enabled)
if events_listeners is not None:
pulumi.set(__self__, "events_listeners", events_listeners)
if id is not None:
pulumi.set(__self__, "id", id)
if identity_providers is not None:
pulumi.set(__self__, "identity_providers", identity_providers)
if users is not None:
pulumi.set(__self__, "users", users)
@property
@pulumi.getter
def realm(self) -> str:
"""
Realm name.
"""
return pulumi.get(self, "realm")
@property
@pulumi.getter(name="adminEventsDetailsEnabled")
def admin_events_details_enabled(self) -> Optional[bool]:
"""
Enable admin events details TODO: change to values and use kubebuilder default annotation once supported
"""
return pulumi.get(self, "admin_events_details_enabled")
@property
@pulumi.getter(name="adminEventsEnabled")
def admin_events_enabled(self) -> Optional[bool]:
"""
Enable events recording TODO: change to values and use kubebuilder default annotation once supported
"""
return pulumi.get(self, "admin_events_enabled")
@property
@pulumi.getter
def clients(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmClients']]:
"""
A set of Keycloak Clients.
"""
return pulumi.get(self, "clients")
@property
@pulumi.getter(name="displayName")
def display_name(self) -> Optional[str]:
"""
Realm display name.
"""
return pulumi.get(self, "display_name")
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
Realm enabled flag.
"""
return pulumi.get(self, "enabled")
@property
@pulumi.getter(name="eventsEnabled")
def events_enabled(self) -> Optional[bool]:
"""
Enable events recording TODO: change to values and use kubebuilder default annotation once supported
"""
return pulumi.get(self, "events_enabled")
@property
@pulumi.getter(name="eventsListeners")
def events_listeners(self) -> Optional[Sequence[str]]:
"""
A set of Event Listeners.
"""
return pulumi.get(self, "events_listeners")
@property
@pulumi.getter
def id(self) -> Optional[str]:
return pulumi.get(self, "id")
@property
@pulumi.getter(name="identityProviders")
def identity_providers(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmIdentityProviders']]:
"""
A set of Identity Providers.
"""
return pulumi.get(self, "identity_providers")
@property
@pulumi.getter
def users(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmUsers']]:
"""
A set of Keycloak Users.
"""
return pulumi.get(self, "users")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecRealmClients(dict):
def __init__(__self__, *,
client_id: str,
access: Optional[Mapping[str, bool]] = None,
admin_url: Optional[str] = None,
attributes: Optional[Mapping[str, str]] = None,
base_url: Optional[str] = None,
bearer_only: Optional[bool] = None,
client_authenticator_type: Optional[str] = None,
consent_required: Optional[bool] = None,
default_roles: Optional[Sequence[str]] = None,
description: Optional[str] = None,
direct_access_grants_enabled: Optional[bool] = None,
enabled: Optional[bool] = None,
frontchannel_logout: Optional[bool] = None,
full_scope_allowed: Optional[bool] = None,
id: Optional[str] = None,
implicit_flow_enabled: Optional[bool] = None,
name: Optional[str] = None,
node_re_registration_timeout: Optional[int] = None,
not_before: Optional[int] = None,
protocol: Optional[str] = None,
protocol_mappers: Optional[Sequence['outputs.KeycloakRealmSpecRealmClientsProtocolMappers']] = None,
public_client: Optional[bool] = None,
redirect_uris: Optional[Sequence[str]] = None,
root_url: Optional[str] = None,
secret: Optional[str] = None,
service_accounts_enabled: Optional[bool] = None,
standard_flow_enabled: Optional[bool] = None,
surrogate_auth_required: Optional[bool] = None,
use_template_config: Optional[bool] = None,
use_template_mappers: Optional[bool] = None,
use_template_scope: Optional[bool] = None,
web_origins: Optional[Sequence[str]] = None):
"""
:param str client_id: Client ID.
:param Mapping[str, bool] access: Access options.
:param str admin_url: Application Admin URL.
:param Mapping[str, str] attributes: Client Attributes.
:param str base_url: Application base URL.
:param bool bearer_only: True if a client supports only Bearer Tokens.
:param str client_authenticator_type: What Client authentication type to use.
:param bool consent_required: True if Consent Screen is required.
:param Sequence[str] default_roles: Default Client roles.
:param str description: Client description.
:param bool direct_access_grants_enabled: True if Direct Grant is enabled.
:param bool enabled: Client enabled flag.
:param bool frontchannel_logout: True if this client supports Front Channel logout.
:param bool full_scope_allowed: True if Full Scope is allowed.
:param str id: Client ID. If not specified, automatically generated.
:param bool implicit_flow_enabled: True if Implicit flow is enabled.
:param str name: Client name.
:param int node_re_registration_timeout: Node registration timeout.
:param int not_before: Not Before setting.
:param str protocol: Protocol used for this Client.
:param Sequence['KeycloakRealmSpecRealmClientsProtocolMappersArgs'] protocol_mappers: Protocol Mappers.
:param bool public_client: True if this is a public Client.
:param Sequence[str] redirect_uris: A list of valid Redirection URLs.
:param str root_url: Application root URL.
:param str secret: Client Secret. The Operator will automatically create a Secret based on this value.
:param bool service_accounts_enabled: True if Service Accounts are enabled.
:param bool standard_flow_enabled: True if Standard flow is enabled.
:param bool surrogate_auth_required: Surrogate Authentication Required option.
:param bool use_template_config: True to use a Template Config.
:param bool use_template_mappers: True to use Template Mappers.
:param bool use_template_scope: True to use Template Scope.
:param Sequence[str] web_origins: A list of valid Web Origins.
"""
pulumi.set(__self__, "client_id", client_id)
if access is not None:
pulumi.set(__self__, "access", access)
if admin_url is not None:
pulumi.set(__self__, "admin_url", admin_url)
if attributes is not None:
pulumi.set(__self__, "attributes", attributes)
if base_url is not None:
pulumi.set(__self__, "base_url", base_url)
if bearer_only is not None:
pulumi.set(__self__, "bearer_only", bearer_only)
if client_authenticator_type is not None:
pulumi.set(__self__, "client_authenticator_type", client_authenticator_type)
if consent_required is not None:
pulumi.set(__self__, "consent_required", consent_required)
if default_roles is not None:
pulumi.set(__self__, "default_roles", default_roles)
if description is not None:
pulumi.set(__self__, "description", description)
if direct_access_grants_enabled is not None:
pulumi.set(__self__, "direct_access_grants_enabled", direct_access_grants_enabled)
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
if frontchannel_logout is not None:
pulumi.set(__self__, "frontchannel_logout", frontchannel_logout)
if full_scope_allowed is not None:
pulumi.set(__self__, "full_scope_allowed", full_scope_allowed)
if id is not None:
pulumi.set(__self__, "id", id)
if implicit_flow_enabled is not None:
pulumi.set(__self__, "implicit_flow_enabled", implicit_flow_enabled)
if name is not None:
pulumi.set(__self__, "name", name)
if node_re_registration_timeout is not None:
pulumi.set(__self__, "node_re_registration_timeout", node_re_registration_timeout)
if not_before is not None:
pulumi.set(__self__, "not_before", not_before)
if protocol is not None:
pulumi.set(__self__, "protocol", protocol)
if protocol_mappers is not None:
pulumi.set(__self__, "protocol_mappers", protocol_mappers)
if public_client is not None:
pulumi.set(__self__, "public_client", public_client)
if redirect_uris is not None:
pulumi.set(__self__, "redirect_uris", redirect_uris)
if root_url is not None:
pulumi.set(__self__, "root_url", root_url)
if secret is not None:
pulumi.set(__self__, "secret", secret)
if service_accounts_enabled is not None:
pulumi.set(__self__, "service_accounts_enabled", service_accounts_enabled)
if standard_flow_enabled is not None:
pulumi.set(__self__, "standard_flow_enabled", standard_flow_enabled)
if surrogate_auth_required is not None:
pulumi.set(__self__, "surrogate_auth_required", surrogate_auth_required)
if use_template_config is not None:
pulumi.set(__self__, "use_template_config", use_template_config)
if use_template_mappers is not None:
pulumi.set(__self__, "use_template_mappers", use_template_mappers)
if use_template_scope is not None:
pulumi.set(__self__, "use_template_scope", use_template_scope)
if web_origins is not None:
pulumi.set(__self__, "web_origins", web_origins)
@property
@pulumi.getter(name="clientId")
def client_id(self) -> str:
"""
Client ID.
"""
return pulumi.get(self, "client_id")
@property
@pulumi.getter
def access(self) -> Optional[Mapping[str, bool]]:
"""
Access options.
"""
return pulumi.get(self, "access")
@property
@pulumi.getter(name="adminUrl")
def admin_url(self) -> Optional[str]:
"""
Application Admin URL.
"""
return pulumi.get(self, "admin_url")
@property
@pulumi.getter
def attributes(self) -> Optional[Mapping[str, str]]:
"""
Client Attributes.
"""
return pulumi.get(self, "attributes")
@property
@pulumi.getter(name="baseUrl")
def base_url(self) -> Optional[str]:
"""
Application base URL.
"""
return pulumi.get(self, "base_url")
@property
@pulumi.getter(name="bearerOnly")
def bearer_only(self) -> Optional[bool]:
"""
True if a client supports only Bearer Tokens.
"""
return pulumi.get(self, "bearer_only")
@property
@pulumi.getter(name="clientAuthenticatorType")
def client_authenticator_type(self) -> Optional[str]:
"""
What Client authentication type to use.
"""
return pulumi.get(self, "client_authenticator_type")
@property
@pulumi.getter(name="consentRequired")
def consent_required(self) -> Optional[bool]:
"""
True if Consent Screen is required.
"""
return pulumi.get(self, "consent_required")
@property
@pulumi.getter(name="defaultRoles")
def default_roles(self) -> Optional[Sequence[str]]:
"""
Default Client roles.
"""
return pulumi.get(self, "default_roles")
@property
@pulumi.getter
def description(self) -> Optional[str]:
"""
Client description.
"""
return pulumi.get(self, "description")
@property
@pulumi.getter(name="directAccessGrantsEnabled")
def direct_access_grants_enabled(self) -> Optional[bool]:
"""
True if Direct Grant is enabled.
"""
return pulumi.get(self, "direct_access_grants_enabled")
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
Client enabled flag.
"""
return pulumi.get(self, "enabled")
@property
@pulumi.getter(name="frontchannelLogout")
def frontchannel_logout(self) -> Optional[bool]:
"""
True if this client supports Front Channel logout.
"""
return pulumi.get(self, "frontchannel_logout")
@property
@pulumi.getter(name="fullScopeAllowed")
def full_scope_allowed(self) -> Optional[bool]:
"""
True if Full Scope is allowed.
"""
return pulumi.get(self, "full_scope_allowed")
@property
@pulumi.getter
def id(self) -> Optional[str]:
"""
Client ID. If not specified, automatically generated.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter(name="implicitFlowEnabled")
def implicit_flow_enabled(self) -> Optional[bool]:
"""
True if Implicit flow is enabled.
"""
return pulumi.get(self, "implicit_flow_enabled")
@property
@pulumi.getter
def name(self) -> Optional[str]:
"""
Client name.
"""
return pulumi.get(self, "name")
@property
@pulumi.getter(name="nodeReRegistrationTimeout")
def node_re_registration_timeout(self) -> Optional[int]:
"""
Node registration timeout.
"""
return pulumi.get(self, "node_re_registration_timeout")
@property
@pulumi.getter(name="notBefore")
def not_before(self) -> Optional[int]:
"""
Not Before setting.
"""
return pulumi.get(self, "not_before")
@property
@pulumi.getter
def protocol(self) -> Optional[str]:
"""
Protocol used for this Client.
"""
return pulumi.get(self, "protocol")
@property
@pulumi.getter(name="protocolMappers")
def protocol_mappers(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmClientsProtocolMappers']]:
"""
Protocol Mappers.
"""
return pulumi.get(self, "protocol_mappers")
@property
@pulumi.getter(name="publicClient")
def public_client(self) -> Optional[bool]:
"""
True if this is a public Client.
"""
return pulumi.get(self, "public_client")
@property
@pulumi.getter(name="redirectUris")
def redirect_uris(self) -> Optional[Sequence[str]]:
"""
A list of valid Redirection URLs.
"""
return pulumi.get(self, "redirect_uris")
@property
@pulumi.getter(name="rootUrl")
def root_url(self) -> Optional[str]:
"""
Application root URL.
"""
return pulumi.get(self, "root_url")
@property
@pulumi.getter
def secret(self) -> Optional[str]:
"""
Client Secret. The Operator will automatically create a Secret based on this value.
"""
return pulumi.get(self, "secret")
@property
@pulumi.getter(name="serviceAccountsEnabled")
def service_accounts_enabled(self) -> Optional[bool]:
"""
True if Service Accounts are enabled.
"""
return pulumi.get(self, "service_accounts_enabled")
@property
@pulumi.getter(name="standardFlowEnabled")
def standard_flow_enabled(self) -> Optional[bool]:
"""
True if Standard flow is enabled.
"""
return pulumi.get(self, "standard_flow_enabled")
@property
@pulumi.getter(name="surrogateAuthRequired")
def surrogate_auth_required(self) -> Optional[bool]:
"""
Surrogate Authentication Required option.
"""
return pulumi.get(self, "surrogate_auth_required")
@property
@pulumi.getter(name="useTemplateConfig")
def use_template_config(self) -> Optional[bool]:
"""
True to use a Template Config.
"""
return pulumi.get(self, "use_template_config")
@property
@pulumi.getter(name="useTemplateMappers")
def use_template_mappers(self) -> Optional[bool]:
"""
True to use Template Mappers.
"""
return pulumi.get(self, "use_template_mappers")
@property
@pulumi.getter(name="useTemplateScope")
def use_template_scope(self) -> Optional[bool]:
"""
True to use Template Scope.
"""
return pulumi.get(self, "use_template_scope")
@property
@pulumi.getter(name="webOrigins")
def web_origins(self) -> Optional[Sequence[str]]:
"""
A list of valid Web Origins.
"""
return pulumi.get(self, "web_origins")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecRealmClientsProtocolMappers(dict):
def __init__(__self__, *,
config: Optional[Mapping[str, str]] = None,
consent_required: Optional[bool] = None,
consent_text: Optional[str] = None,
id: Optional[str] = None,
name: Optional[str] = None,
protocol: Optional[str] = None,
protocol_mapper: Optional[str] = None):
"""
:param Mapping[str, str] config: Config options.
:param bool consent_required: True if Consent Screen is required.
:param str consent_text: Text to use for displaying Consent Screen.
:param str id: Protocol Mapper ID.
:param str name: Protocol Mapper Name.
:param str protocol: Protocol to use.
:param str protocol_mapper: Protocol Mapper to use
"""
if config is not None:
pulumi.set(__self__, "config", config)
if consent_required is not None:
pulumi.set(__self__, "consent_required", consent_required)
if consent_text is not None:
pulumi.set(__self__, "consent_text", consent_text)
if id is not None:
pulumi.set(__self__, "id", id)
if name is not None:
pulumi.set(__self__, "name", name)
if protocol is not None:
pulumi.set(__self__, "protocol", protocol)
if protocol_mapper is not None:
pulumi.set(__self__, "protocol_mapper", protocol_mapper)
@property
@pulumi.getter
def config(self) -> Optional[Mapping[str, str]]:
"""
Config options.
"""
return pulumi.get(self, "config")
@property
@pulumi.getter(name="consentRequired")
def consent_required(self) -> Optional[bool]:
"""
True if Consent Screen is required.
"""
return pulumi.get(self, "consent_required")
@property
@pulumi.getter(name="consentText")
def consent_text(self) -> Optional[str]:
"""
Text to use for displaying Consent Screen.
"""
return pulumi.get(self, "consent_text")
@property
@pulumi.getter
def id(self) -> Optional[str]:
"""
Protocol Mapper ID.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter
def name(self) -> Optional[str]:
"""
Protocol Mapper Name.
"""
return pulumi.get(self, "name")
@property
@pulumi.getter
def protocol(self) -> Optional[str]:
"""
Protocol to use.
"""
return pulumi.get(self, "protocol")
@property
@pulumi.getter(name="protocolMapper")
def protocol_mapper(self) -> Optional[str]:
"""
Protocol Mapper to use
"""
return pulumi.get(self, "protocol_mapper")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecRealmIdentityProviders(dict):
def __init__(__self__, *,
add_read_token_role_on_create: Optional[bool] = None,
alias: Optional[str] = None,
config: Optional[Mapping[str, str]] = None,
display_name: Optional[str] = None,
enabled: Optional[bool] = None,
first_broker_login_flow_alias: Optional[str] = None,
internal_id: Optional[str] = None,
link_only: Optional[bool] = None,
post_broker_login_flow_alias: Optional[str] = None,
provider_id: Optional[str] = None,
store_token: Optional[bool] = None,
trust_email: Optional[bool] = None):
"""
:param bool add_read_token_role_on_create: Adds Read Token role when creating this Identity Provider.
:param str alias: Identity Provider Alias.
:param Mapping[str, str] config: Identity Provider config.
:param str display_name: Identity Provider Display Name.
:param bool enabled: Identity Provider enabled flag.
:param str first_broker_login_flow_alias: Identity Provider First Broker Login Flow Alias.
:param str internal_id: Identity Provider Internal ID.
:param bool link_only: Identity Provider Link Only setting.
:param str post_broker_login_flow_alias: Identity Provider Post Broker Login Flow Alias.
:param str provider_id: Identity Provider ID.
:param bool store_token: Identity Provider Store to Token.
:param bool trust_email: Identity Provider Trust Email.
"""
if add_read_token_role_on_create is not None:
pulumi.set(__self__, "add_read_token_role_on_create", add_read_token_role_on_create)
if alias is not None:
pulumi.set(__self__, "alias", alias)
if config is not None:
pulumi.set(__self__, "config", config)
if display_name is not None:
pulumi.set(__self__, "display_name", display_name)
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
if first_broker_login_flow_alias is not None:
pulumi.set(__self__, "first_broker_login_flow_alias", first_broker_login_flow_alias)
if internal_id is not None:
pulumi.set(__self__, "internal_id", internal_id)
if link_only is not None:
pulumi.set(__self__, "link_only", link_only)
if post_broker_login_flow_alias is not None:
pulumi.set(__self__, "post_broker_login_flow_alias", post_broker_login_flow_alias)
if provider_id is not None:
pulumi.set(__self__, "provider_id", provider_id)
if store_token is not None:
pulumi.set(__self__, "store_token", store_token)
if trust_email is not None:
pulumi.set(__self__, "trust_email", trust_email)
@property
@pulumi.getter(name="addReadTokenRoleOnCreate")
def add_read_token_role_on_create(self) -> Optional[bool]:
"""
Adds Read Token role when creating this Identity Provider.
"""
return pulumi.get(self, "add_read_token_role_on_create")
@property
@pulumi.getter
def alias(self) -> Optional[str]:
"""
Identity Provider Alias.
"""
return pulumi.get(self, "alias")
@property
@pulumi.getter
def config(self) -> Optional[Mapping[str, str]]:
"""
Identity Provider config.
"""
return pulumi.get(self, "config")
@property
@pulumi.getter(name="displayName")
def display_name(self) -> Optional[str]:
"""
Identity Provider Display Name.
"""
return pulumi.get(self, "display_name")
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
Identity Provider enabled flag.
"""
return pulumi.get(self, "enabled")
@property
@pulumi.getter(name="firstBrokerLoginFlowAlias")
def first_broker_login_flow_alias(self) -> Optional[str]:
"""
Identity Provider First Broker Login Flow Alias.
"""
return pulumi.get(self, "first_broker_login_flow_alias")
@property
@pulumi.getter(name="internalId")
def internal_id(self) -> Optional[str]:
"""
Identity Provider Internal ID.
"""
return pulumi.get(self, "internal_id")
@property
@pulumi.getter(name="linkOnly")
def link_only(self) -> Optional[bool]:
"""
Identity Provider Link Only setting.
"""
return pulumi.get(self, "link_only")
@property
@pulumi.getter(name="postBrokerLoginFlowAlias")
def post_broker_login_flow_alias(self) -> Optional[str]:
"""
Identity Provider Post Broker Login Flow Alias.
"""
return pulumi.get(self, "post_broker_login_flow_alias")
@property
@pulumi.getter(name="providerId")
def provider_id(self) -> Optional[str]:
"""
Identity Provider ID.
"""
return pulumi.get(self, "provider_id")
@property
@pulumi.getter(name="storeToken")
def store_token(self) -> Optional[bool]:
"""
Identity Provider Store to Token.
"""
return pulumi.get(self, "store_token")
@property
@pulumi.getter(name="trustEmail")
def trust_email(self) -> Optional[bool]:
"""
Identity Provider Trust Email.
"""
return pulumi.get(self, "trust_email")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecRealmOverrides(dict):
def __init__(__self__, *,
for_flow: Optional[str] = None,
identity_provider: Optional[str] = None):
"""
:param str for_flow: Flow to be overridden.
:param str identity_provider: Identity Provider to be overridden.
"""
if for_flow is not None:
pulumi.set(__self__, "for_flow", for_flow)
if identity_provider is not None:
pulumi.set(__self__, "identity_provider", identity_provider)
@property
@pulumi.getter(name="forFlow")
def for_flow(self) -> Optional[str]:
"""
Flow to be overridden.
"""
return pulumi.get(self, "for_flow")
@property
@pulumi.getter(name="identityProvider")
def identity_provider(self) -> Optional[str]:
"""
Identity Provider to be overridden.
"""
return pulumi.get(self, "identity_provider")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecRealmUsers(dict):
def __init__(__self__, *,
attributes: Optional[Mapping[str, Sequence[str]]] = None,
client_roles: Optional[Mapping[str, Sequence[str]]] = None,
credentials: Optional[Sequence['outputs.KeycloakRealmSpecRealmUsersCredentials']] = None,
email: Optional[str] = None,
email_verified: Optional[bool] = None,
enabled: Optional[bool] = None,
federated_identities: Optional[Sequence['outputs.KeycloakRealmSpecRealmUsersFederatedIdentities']] = None,
first_name: Optional[str] = None,
groups: Optional[Sequence[str]] = None,
id: Optional[str] = None,
last_name: Optional[str] = None,
realm_roles: Optional[Sequence[str]] = None,
required_actions: Optional[Sequence[str]] = None,
username: Optional[str] = None):
"""
:param Mapping[str, Sequence[str]] attributes: A set of Attributes.
:param Mapping[str, Sequence[str]] client_roles: A set of Client Roles.
:param Sequence['KeycloakRealmSpecRealmUsersCredentialsArgs'] credentials: A set of Credentials.
:param str email: Email.
:param bool email_verified: True if email has already been verified.
:param bool enabled: User enabled flag.
:param Sequence['KeycloakRealmSpecRealmUsersFederatedIdentitiesArgs'] federated_identities: A set of Federated Identities.
:param str first_name: First Name.
:param Sequence[str] groups: A set of Groups.
:param str id: User ID.
:param str last_name: Last Name.
:param Sequence[str] realm_roles: A set of Realm Roles.
:param Sequence[str] required_actions: A set of Required Actions.
:param str username: User Name.
"""
if attributes is not None:
pulumi.set(__self__, "attributes", attributes)
if client_roles is not None:
pulumi.set(__self__, "client_roles", client_roles)
if credentials is not None:
pulumi.set(__self__, "credentials", credentials)
if email is not None:
pulumi.set(__self__, "email", email)
if email_verified is not None:
pulumi.set(__self__, "email_verified", email_verified)
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
if federated_identities is not None:
pulumi.set(__self__, "federated_identities", federated_identities)
if first_name is not None:
pulumi.set(__self__, "first_name", first_name)
if groups is not None:
pulumi.set(__self__, "groups", groups)
if id is not None:
pulumi.set(__self__, "id", id)
if last_name is not None:
pulumi.set(__self__, "last_name", last_name)
if realm_roles is not None:
pulumi.set(__self__, "realm_roles", realm_roles)
if required_actions is not None:
pulumi.set(__self__, "required_actions", required_actions)
if username is not None:
pulumi.set(__self__, "username", username)
@property
@pulumi.getter
def attributes(self) -> Optional[Mapping[str, Sequence[str]]]:
"""
A set of Attributes.
"""
return pulumi.get(self, "attributes")
@property
@pulumi.getter(name="clientRoles")
def client_roles(self) -> Optional[Mapping[str, Sequence[str]]]:
"""
A set of Client Roles.
"""
return pulumi.get(self, "client_roles")
@property
@pulumi.getter
def credentials(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmUsersCredentials']]:
"""
A set of Credentials.
"""
return pulumi.get(self, "credentials")
@property
@pulumi.getter
def email(self) -> Optional[str]:
"""
Email.
"""
return pulumi.get(self, "email")
@property
@pulumi.getter(name="emailVerified")
def email_verified(self) -> Optional[bool]:
"""
True if email has already been verified.
"""
return pulumi.get(self, "email_verified")
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
User enabled flag.
"""
return pulumi.get(self, "enabled")
@property
@pulumi.getter(name="federatedIdentities")
def federated_identities(self) -> Optional[Sequence['outputs.KeycloakRealmSpecRealmUsersFederatedIdentities']]:
"""
A set of Federated Identities.
"""
return pulumi.get(self, "federated_identities")
@property
@pulumi.getter(name="firstName")
def first_name(self) -> Optional[str]:
"""
First Name.
"""
return pulumi.get(self, "first_name")
@property
@pulumi.getter
def groups(self) -> Optional[Sequence[str]]:
"""
A set of Groups.
"""
return pulumi.get(self, "groups")
@property
@pulumi.getter
def id(self) -> Optional[str]:
"""
User ID.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter(name="lastName")
def last_name(self) -> Optional[str]:
"""
Last Name.
"""
return pulumi.get(self, "last_name")
@property
@pulumi.getter(name="realmRoles")
def realm_roles(self) -> Optional[Sequence[str]]:
"""
A set of Realm Roles.
"""
return pulumi.get(self, "realm_roles")
@property
@pulumi.getter(name="requiredActions")
def required_actions(self) -> Optional[Sequence[str]]:
"""
A set of Required Actions.
"""
return pulumi.get(self, "required_actions")
@property
@pulumi.getter
def username(self) -> Optional[str]:
"""
User Name.
"""
return pulumi.get(self, "username")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecRealmUsersCredentials(dict):
def __init__(__self__, *,
temporary: Optional[bool] = None,
type: Optional[str] = None,
value: Optional[str] = None):
"""
:param bool temporary: True if this credential object is temporary.
:param str type: Credential Type.
:param str value: Credential Value.
"""
if temporary is not None:
pulumi.set(__self__, "temporary", temporary)
if type is not None:
pulumi.set(__self__, "type", type)
if value is not None:
pulumi.set(__self__, "value", value)
@property
@pulumi.getter
def temporary(self) -> Optional[bool]:
"""
True if this credential object is temporary.
"""
return pulumi.get(self, "temporary")
@property
@pulumi.getter
def type(self) -> Optional[str]:
"""
Credential Type.
"""
return pulumi.get(self, "type")
@property
@pulumi.getter
def value(self) -> Optional[str]:
"""
Credential Value.
"""
return pulumi.get(self, "value")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmSpecRealmUsersFederatedIdentities(dict):
def __init__(__self__, *,
identity_provider: Optional[str] = None,
user_id: Optional[str] = None,
user_name: Optional[str] = None):
"""
:param str identity_provider: Federated Identity Provider.
:param str user_id: Federated Identity User ID.
:param str user_name: Federated Identity User Name.
"""
if identity_provider is not None:
pulumi.set(__self__, "identity_provider", identity_provider)
if user_id is not None:
pulumi.set(__self__, "user_id", user_id)
if user_name is not None:
pulumi.set(__self__, "user_name", user_name)
@property
@pulumi.getter(name="identityProvider")
def identity_provider(self) -> Optional[str]:
"""
Federated Identity Provider.
"""
return pulumi.get(self, "identity_provider")
@property
@pulumi.getter(name="userId")
def user_id(self) -> Optional[str]:
"""
Federated Identity User ID.
"""
return pulumi.get(self, "user_id")
@property
@pulumi.getter(name="userName")
def user_name(self) -> Optional[str]:
"""
Federated Identity User Name.
"""
return pulumi.get(self, "user_name")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakRealmStatus(dict):
"""
KeycloakRealmStatus defines the observed state of KeycloakRealm
"""
def __init__(__self__, *,
login_url: str,
message: str,
phase: str,
ready: bool,
secondary_resources: Optional[Mapping[str, Sequence[str]]] = None):
"""
KeycloakRealmStatus defines the observed state of KeycloakRealm
:param str login_url: TODO
:param str message: Human-readable message indicating details about current operator phase or error.
:param str phase: Current phase of the operator.
:param bool ready: True if all resources are in a ready state and all work is done.
:param Mapping[str, Sequence[str]] secondary_resources: A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ]
"""
pulumi.set(__self__, "login_url", login_url)
pulumi.set(__self__, "message", message)
pulumi.set(__self__, "phase", phase)
pulumi.set(__self__, "ready", ready)
if secondary_resources is not None:
pulumi.set(__self__, "secondary_resources", secondary_resources)
@property
@pulumi.getter(name="loginURL")
def login_url(self) -> str:
"""
TODO
"""
return pulumi.get(self, "login_url")
@property
@pulumi.getter
def message(self) -> str:
"""
Human-readable message indicating details about current operator phase or error.
"""
return pulumi.get(self, "message")
@property
@pulumi.getter
def phase(self) -> str:
"""
Current phase of the operator.
"""
return pulumi.get(self, "phase")
@property
@pulumi.getter
def ready(self) -> bool:
"""
True if all resources are in a ready state and all work is done.
"""
return pulumi.get(self, "ready")
@property
@pulumi.getter(name="secondaryResources")
def secondary_resources(self) -> Optional[Mapping[str, Sequence[str]]]:
"""
A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ]
"""
return pulumi.get(self, "secondary_resources")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpec(dict):
"""
KeycloakSpec defines the desired state of Keycloak.
"""
def __init__(__self__, *,
extensions: Optional[Sequence[str]] = None,
external_access: Optional['outputs.KeycloakSpecExternalAccess'] = None,
external_database: Optional['outputs.KeycloakSpecExternalDatabase'] = None,
instances: Optional[int] = None,
keycloak_deployment_spec: Optional['outputs.KeycloakSpecKeycloakDeploymentSpec'] = None,
migration: Optional['outputs.KeycloakSpecMigration'] = None,
pod_disruption_budget: Optional['outputs.KeycloakSpecPodDisruptionBudget'] = None,
postgres_deployment_spec: Optional['outputs.KeycloakSpecPostgresDeploymentSpec'] = None,
profile: Optional[str] = None,
storage_class_name: Optional[str] = None):
"""
KeycloakSpec defines the desired state of Keycloak.
:param Sequence[str] extensions: A list of extensions, where each one is a URL to a JAR files that will be deployed in Keycloak.
:param 'KeycloakSpecExternalAccessArgs' external_access: Controls external Ingress/Route settings.
:param 'KeycloakSpecExternalDatabaseArgs' external_database: Controls external database settings. Using an external database requires providing a secret containing credentials as well as connection details. Here's an example of such secret:
apiVersion: v1 kind: Secret metadata: name: keycloak-db-secret namespace: keycloak stringData: POSTGRES_DATABASE: <Database Name> POSTGRES_EXTERNAL_ADDRESS: <External Database IP or URL (resolvable by K8s)> POSTGRES_EXTERNAL_PORT: <External Database Port> # Strongly recommended to use <'Keycloak CR Name'-postgresql> POSTGRES_HOST: <Database Service Name> POSTGRES_PASSWORD: <Database Password> # Required for AWS Backup functionality POSTGRES_SUPERUSER: true POSTGRES_USERNAME: <Database Username> type: Opaque
Both POSTGRES_EXTERNAL_ADDRESS and POSTGRES_EXTERNAL_PORT are specifically required for creating connection to the external database. The secret name is created using the following convention: <Custom Resource Name>-db-secret
For more information, please refer to the Operator documentation.
:param int instances: Number of Keycloak instances in HA mode. Default is 1.
:param 'KeycloakSpecKeycloakDeploymentSpecArgs' keycloak_deployment_spec: Resources (Requests and Limits) for KeycloakDeployment.
:param 'KeycloakSpecMigrationArgs' migration: Specify Migration configuration
:param 'KeycloakSpecPodDisruptionBudgetArgs' pod_disruption_budget: Specify PodDisruptionBudget configuration.
:param 'KeycloakSpecPostgresDeploymentSpecArgs' postgres_deployment_spec: Resources (Requests and Limits) for PostgresDeployment.
:param str profile: Profile used for controlling Operator behavior. Default is empty.
:param str storage_class_name: Name of the StorageClass for Postgresql Persistent Volume Claim
"""
if extensions is not None:
pulumi.set(__self__, "extensions", extensions)
if external_access is not None:
pulumi.set(__self__, "external_access", external_access)
if external_database is not None:
pulumi.set(__self__, "external_database", external_database)
if instances is not None:
pulumi.set(__self__, "instances", instances)
if keycloak_deployment_spec is not None:
pulumi.set(__self__, "keycloak_deployment_spec", keycloak_deployment_spec)
if migration is not None:
pulumi.set(__self__, "migration", migration)
if pod_disruption_budget is not None:
pulumi.set(__self__, "pod_disruption_budget", pod_disruption_budget)
if postgres_deployment_spec is not None:
pulumi.set(__self__, "postgres_deployment_spec", postgres_deployment_spec)
if profile is not None:
pulumi.set(__self__, "profile", profile)
if storage_class_name is not None:
pulumi.set(__self__, "storage_class_name", storage_class_name)
@property
@pulumi.getter
def extensions(self) -> Optional[Sequence[str]]:
"""
A list of extensions, where each one is a URL to a JAR files that will be deployed in Keycloak.
"""
return pulumi.get(self, "extensions")
@property
@pulumi.getter(name="externalAccess")
def external_access(self) -> Optional['outputs.KeycloakSpecExternalAccess']:
"""
Controls external Ingress/Route settings.
"""
return pulumi.get(self, "external_access")
@property
@pulumi.getter(name="externalDatabase")
def external_database(self) -> Optional['outputs.KeycloakSpecExternalDatabase']:
"""
Controls external database settings. Using an external database requires providing a secret containing credentials as well as connection details. Here's an example of such secret:
apiVersion: v1 kind: Secret metadata: name: keycloak-db-secret namespace: keycloak stringData: POSTGRES_DATABASE: <Database Name> POSTGRES_EXTERNAL_ADDRESS: <External Database IP or URL (resolvable by K8s)> POSTGRES_EXTERNAL_PORT: <External Database Port> # Strongly recommended to use <'Keycloak CR Name'-postgresql> POSTGRES_HOST: <Database Service Name> POSTGRES_PASSWORD: <Database Password> # Required for AWS Backup functionality POSTGRES_SUPERUSER: true POSTGRES_USERNAME: <Database Username> type: Opaque
Both POSTGRES_EXTERNAL_ADDRESS and POSTGRES_EXTERNAL_PORT are specifically required for creating connection to the external database. The secret name is created using the following convention: <Custom Resource Name>-db-secret
For more information, please refer to the Operator documentation.
"""
return pulumi.get(self, "external_database")
@property
@pulumi.getter
def instances(self) -> Optional[int]:
"""
Number of Keycloak instances in HA mode. Default is 1.
"""
return pulumi.get(self, "instances")
@property
@pulumi.getter(name="keycloakDeploymentSpec")
def keycloak_deployment_spec(self) -> Optional['outputs.KeycloakSpecKeycloakDeploymentSpec']:
"""
Resources (Requests and Limits) for KeycloakDeployment.
"""
return pulumi.get(self, "keycloak_deployment_spec")
@property
@pulumi.getter
def migration(self) -> Optional['outputs.KeycloakSpecMigration']:
"""
Specify Migration configuration
"""
return pulumi.get(self, "migration")
@property
@pulumi.getter(name="podDisruptionBudget")
def pod_disruption_budget(self) -> Optional['outputs.KeycloakSpecPodDisruptionBudget']:
"""
Specify PodDisruptionBudget configuration.
"""
return pulumi.get(self, "pod_disruption_budget")
@property
@pulumi.getter(name="postgresDeploymentSpec")
def postgres_deployment_spec(self) -> Optional['outputs.KeycloakSpecPostgresDeploymentSpec']:
"""
Resources (Requests and Limits) for PostgresDeployment.
"""
return pulumi.get(self, "postgres_deployment_spec")
@property
@pulumi.getter
def profile(self) -> Optional[str]:
"""
Profile used for controlling Operator behavior. Default is empty.
"""
return pulumi.get(self, "profile")
@property
@pulumi.getter(name="storageClassName")
def storage_class_name(self) -> Optional[str]:
"""
Name of the StorageClass for Postgresql Persistent Volume Claim
"""
return pulumi.get(self, "storage_class_name")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecExternalAccess(dict):
"""
Controls external Ingress/Route settings.
"""
def __init__(__self__, *,
enabled: Optional[bool] = None,
tls_termination: Optional[str] = None):
"""
Controls external Ingress/Route settings.
:param bool enabled: If set to true, the Operator will create an Ingress or a Route pointing to Keycloak.
:param str tls_termination: TLS Termination type for the external access. Setting this field to "reencrypt" will terminate TLS on the Ingress/Route level. Setting this field to "passthrough" will send encrypted traffic to the Pod. If unspecified, defaults to "reencrypt". Note, that this setting has no effect on Ingress as Ingress TLS settings are not reconciled by this operator. In other words, Ingress TLS configuration is the same in both cases and it is up to the user to configure TLS section of the Ingress.
"""
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
if tls_termination is not None:
pulumi.set(__self__, "tls_termination", tls_termination)
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
If set to true, the Operator will create an Ingress or a Route pointing to Keycloak.
"""
return pulumi.get(self, "enabled")
@property
@pulumi.getter(name="tlsTermination")
def tls_termination(self) -> Optional[str]:
"""
TLS Termination type for the external access. Setting this field to "reencrypt" will terminate TLS on the Ingress/Route level. Setting this field to "passthrough" will send encrypted traffic to the Pod. If unspecified, defaults to "reencrypt". Note, that this setting has no effect on Ingress as Ingress TLS settings are not reconciled by this operator. In other words, Ingress TLS configuration is the same in both cases and it is up to the user to configure TLS section of the Ingress.
"""
return pulumi.get(self, "tls_termination")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecExternalDatabase(dict):
"""
Controls external database settings. Using an external database requires providing a secret containing credentials as well as connection details. Here's an example of such secret:
apiVersion: v1 kind: Secret metadata: name: keycloak-db-secret namespace: keycloak stringData: POSTGRES_DATABASE: <Database Name> POSTGRES_EXTERNAL_ADDRESS: <External Database IP or URL (resolvable by K8s)> POSTGRES_EXTERNAL_PORT: <External Database Port> # Strongly recommended to use <'Keycloak CR Name'-postgresql> POSTGRES_HOST: <Database Service Name> POSTGRES_PASSWORD: <Database Password> # Required for AWS Backup functionality POSTGRES_SUPERUSER: true POSTGRES_USERNAME: <Database Username> type: Opaque
Both POSTGRES_EXTERNAL_ADDRESS and POSTGRES_EXTERNAL_PORT are specifically required for creating connection to the external database. The secret name is created using the following convention: <Custom Resource Name>-db-secret
For more information, please refer to the Operator documentation.
"""
def __init__(__self__, *,
enabled: Optional[bool] = None):
"""
Controls external database settings. Using an external database requires providing a secret containing credentials as well as connection details. Here's an example of such secret:
apiVersion: v1 kind: Secret metadata: name: keycloak-db-secret namespace: keycloak stringData: POSTGRES_DATABASE: <Database Name> POSTGRES_EXTERNAL_ADDRESS: <External Database IP or URL (resolvable by K8s)> POSTGRES_EXTERNAL_PORT: <External Database Port> # Strongly recommended to use <'Keycloak CR Name'-postgresql> POSTGRES_HOST: <Database Service Name> POSTGRES_PASSWORD: <Database Password> # Required for AWS Backup functionality POSTGRES_SUPERUSER: true POSTGRES_USERNAME: <Database Username> type: Opaque
Both POSTGRES_EXTERNAL_ADDRESS and POSTGRES_EXTERNAL_PORT are specifically required for creating connection to the external database. The secret name is created using the following convention: <Custom Resource Name>-db-secret
For more information, please refer to the Operator documentation.
:param bool enabled: If set to true, the Operator will use an external database. pointing to Keycloak.
"""
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
If set to true, the Operator will use an external database. pointing to Keycloak.
"""
return pulumi.get(self, "enabled")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecKeycloakDeploymentSpec(dict):
"""
Resources (Requests and Limits) for KeycloakDeployment.
"""
def __init__(__self__, *,
resources: Optional['outputs.KeycloakSpecKeycloakDeploymentSpecResources'] = None):
"""
Resources (Requests and Limits) for KeycloakDeployment.
:param 'KeycloakSpecKeycloakDeploymentSpecResourcesArgs' resources: Resources (Requests and Limits) for the Pods.
"""
if resources is not None:
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> Optional['outputs.KeycloakSpecKeycloakDeploymentSpecResources']:
"""
Resources (Requests and Limits) for the Pods.
"""
return pulumi.get(self, "resources")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecKeycloakDeploymentSpecResources(dict):
"""
Resources (Requests and Limits) for the Pods.
"""
def __init__(__self__, *,
limits: Optional[Mapping[str, str]] = None,
requests: Optional[Mapping[str, str]] = None):
"""
Resources (Requests and Limits) for the Pods.
:param Mapping[str, str] limits: Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
:param Mapping[str, str] requests: Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
"""
if limits is not None:
pulumi.set(__self__, "limits", limits)
if requests is not None:
pulumi.set(__self__, "requests", requests)
@property
@pulumi.getter
def limits(self) -> Optional[Mapping[str, str]]:
"""
Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
"""
return pulumi.get(self, "limits")
@property
@pulumi.getter
def requests(self) -> Optional[Mapping[str, str]]:
"""
Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
"""
return pulumi.get(self, "requests")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecMigration(dict):
"""
Specify Migration configuration
"""
def __init__(__self__, *,
backups: Optional['outputs.KeycloakSpecMigrationBackups'] = None):
"""
Specify Migration configuration
:param 'KeycloakSpecMigrationBackupsArgs' backups: Set it to config backup policy for migration
"""
if backups is not None:
pulumi.set(__self__, "backups", backups)
@property
@pulumi.getter
def backups(self) -> Optional['outputs.KeycloakSpecMigrationBackups']:
"""
Set it to config backup policy for migration
"""
return pulumi.get(self, "backups")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecMigrationBackups(dict):
"""
Set it to config backup policy for migration
"""
def __init__(__self__, *,
enabled: Optional[bool] = None):
"""
Set it to config backup policy for migration
:param bool enabled: If set to true, the operator will do database backup before doing migration
"""
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
If set to true, the operator will do database backup before doing migration
"""
return pulumi.get(self, "enabled")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecPodDisruptionBudget(dict):
"""
Specify PodDisruptionBudget configuration.
"""
def __init__(__self__, *,
enabled: Optional[bool] = None):
"""
Specify PodDisruptionBudget configuration.
:param bool enabled: If set to true, the operator will create a PodDistruptionBudget for the Keycloak deployment and set its `maxUnavailable` value to 1.
"""
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
If set to true, the operator will create a PodDistruptionBudget for the Keycloak deployment and set its `maxUnavailable` value to 1.
"""
return pulumi.get(self, "enabled")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecPostgresDeploymentSpec(dict):
"""
Resources (Requests and Limits) for PostgresDeployment.
"""
def __init__(__self__, *,
resources: Optional['outputs.KeycloakSpecPostgresDeploymentSpecResources'] = None):
"""
Resources (Requests and Limits) for PostgresDeployment.
:param 'KeycloakSpecPostgresDeploymentSpecResourcesArgs' resources: Resources (Requests and Limits) for the Pods.
"""
if resources is not None:
pulumi.set(__self__, "resources", resources)
@property
@pulumi.getter
def resources(self) -> Optional['outputs.KeycloakSpecPostgresDeploymentSpecResources']:
"""
Resources (Requests and Limits) for the Pods.
"""
return pulumi.get(self, "resources")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakSpecPostgresDeploymentSpecResources(dict):
"""
Resources (Requests and Limits) for the Pods.
"""
def __init__(__self__, *,
limits: Optional[Mapping[str, str]] = None,
requests: Optional[Mapping[str, str]] = None):
"""
Resources (Requests and Limits) for the Pods.
:param Mapping[str, str] limits: Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
:param Mapping[str, str] requests: Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
"""
if limits is not None:
pulumi.set(__self__, "limits", limits)
if requests is not None:
pulumi.set(__self__, "requests", requests)
@property
@pulumi.getter
def limits(self) -> Optional[Mapping[str, str]]:
"""
Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
"""
return pulumi.get(self, "limits")
@property
@pulumi.getter
def requests(self) -> Optional[Mapping[str, str]]:
"""
Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
"""
return pulumi.get(self, "requests")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakStatus(dict):
"""
KeycloakStatus defines the observed state of Keycloak.
"""
def __init__(__self__, *,
credential_secret: str,
internal_url: str,
message: str,
phase: str,
ready: bool,
version: str,
secondary_resources: Optional[Mapping[str, Sequence[str]]] = None):
"""
KeycloakStatus defines the observed state of Keycloak.
:param str credential_secret: The secret where the admin credentials are to be found.
:param str internal_url: Service IP and Port for in-cluster access to the keycloak instance.
:param str message: Human-readable message indicating details about current operator phase or error.
:param str phase: Current phase of the operator.
:param bool ready: True if all resources are in a ready state and all work is done.
:param str version: Version of Keycloak or RHSSO running on the cluster.
:param Mapping[str, Sequence[str]] secondary_resources: A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ].
"""
pulumi.set(__self__, "credential_secret", credential_secret)
pulumi.set(__self__, "internal_url", internal_url)
pulumi.set(__self__, "message", message)
pulumi.set(__self__, "phase", phase)
pulumi.set(__self__, "ready", ready)
pulumi.set(__self__, "version", version)
if secondary_resources is not None:
pulumi.set(__self__, "secondary_resources", secondary_resources)
@property
@pulumi.getter(name="credentialSecret")
def credential_secret(self) -> str:
"""
The secret where the admin credentials are to be found.
"""
return pulumi.get(self, "credential_secret")
@property
@pulumi.getter(name="internalURL")
def internal_url(self) -> str:
"""
Service IP and Port for in-cluster access to the keycloak instance.
"""
return pulumi.get(self, "internal_url")
@property
@pulumi.getter
def message(self) -> str:
"""
Human-readable message indicating details about current operator phase or error.
"""
return pulumi.get(self, "message")
@property
@pulumi.getter
def phase(self) -> str:
"""
Current phase of the operator.
"""
return pulumi.get(self, "phase")
@property
@pulumi.getter
def ready(self) -> bool:
"""
True if all resources are in a ready state and all work is done.
"""
return pulumi.get(self, "ready")
@property
@pulumi.getter
def version(self) -> str:
"""
Version of Keycloak or RHSSO running on the cluster.
"""
return pulumi.get(self, "version")
@property
@pulumi.getter(name="secondaryResources")
def secondary_resources(self) -> Optional[Mapping[str, Sequence[str]]]:
"""
A map of all the secondary resources types and names created for this CR. e.g "Deployment": [ "DeploymentName1", "DeploymentName2" ].
"""
return pulumi.get(self, "secondary_resources")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakUserSpec(dict):
"""
KeycloakUserSpec defines the desired state of KeycloakUser.
"""
def __init__(__self__, *,
user: 'outputs.KeycloakUserSpecUser',
realm_selector: Optional['outputs.KeycloakUserSpecRealmSelector'] = None):
"""
KeycloakUserSpec defines the desired state of KeycloakUser.
:param 'KeycloakUserSpecUserArgs' user: Keycloak User REST object.
:param 'KeycloakUserSpecRealmSelectorArgs' realm_selector: Selector for looking up KeycloakRealm Custom Resources.
"""
pulumi.set(__self__, "user", user)
if realm_selector is not None:
pulumi.set(__self__, "realm_selector", realm_selector)
@property
@pulumi.getter
def user(self) -> 'outputs.KeycloakUserSpecUser':
"""
Keycloak User REST object.
"""
return pulumi.get(self, "user")
@property
@pulumi.getter(name="realmSelector")
def realm_selector(self) -> Optional['outputs.KeycloakUserSpecRealmSelector']:
"""
Selector for looking up KeycloakRealm Custom Resources.
"""
return pulumi.get(self, "realm_selector")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakUserSpecRealmSelector(dict):
"""
Selector for looking up KeycloakRealm Custom Resources.
"""
def __init__(__self__, *,
match_expressions: Optional[Sequence['outputs.KeycloakUserSpecRealmSelectorMatchExpressions']] = None,
match_labels: Optional[Mapping[str, str]] = None):
"""
Selector for looking up KeycloakRealm Custom Resources.
:param Sequence['KeycloakUserSpecRealmSelectorMatchExpressionsArgs'] match_expressions: matchExpressions is a list of label selector requirements. The requirements are ANDed.
:param Mapping[str, str] match_labels: matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.
"""
if match_expressions is not None:
pulumi.set(__self__, "match_expressions", match_expressions)
if match_labels is not None:
pulumi.set(__self__, "match_labels", match_labels)
@property
@pulumi.getter(name="matchExpressions")
def match_expressions(self) -> Optional[Sequence['outputs.KeycloakUserSpecRealmSelectorMatchExpressions']]:
"""
matchExpressions is a list of label selector requirements. The requirements are ANDed.
"""
return pulumi.get(self, "match_expressions")
@property
@pulumi.getter(name="matchLabels")
def match_labels(self) -> Optional[Mapping[str, str]]:
"""
matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.
"""
return pulumi.get(self, "match_labels")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakUserSpecRealmSelectorMatchExpressions(dict):
"""
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
"""
def __init__(__self__, *,
key: str,
operator: str,
values: Optional[Sequence[str]] = None):
"""
A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values.
:param str key: key is the label key that the selector applies to.
:param str operator: operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.
:param Sequence[str] values: values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.
"""
pulumi.set(__self__, "key", key)
pulumi.set(__self__, "operator", operator)
if values is not None:
pulumi.set(__self__, "values", values)
@property
@pulumi.getter
def key(self) -> str:
"""
key is the label key that the selector applies to.
"""
return pulumi.get(self, "key")
@property
@pulumi.getter
def operator(self) -> str:
"""
operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.
"""
return pulumi.get(self, "operator")
@property
@pulumi.getter
def values(self) -> Optional[Sequence[str]]:
"""
values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.
"""
return pulumi.get(self, "values")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakUserSpecUser(dict):
"""
Keycloak User REST object.
"""
def __init__(__self__, *,
attributes: Optional[Mapping[str, Sequence[str]]] = None,
client_roles: Optional[Mapping[str, Sequence[str]]] = None,
credentials: Optional[Sequence['outputs.KeycloakUserSpecUserCredentials']] = None,
email: Optional[str] = None,
email_verified: Optional[bool] = None,
enabled: Optional[bool] = None,
federated_identities: Optional[Sequence['outputs.KeycloakUserSpecUserFederatedIdentities']] = None,
first_name: Optional[str] = None,
groups: Optional[Sequence[str]] = None,
id: Optional[str] = None,
last_name: Optional[str] = None,
realm_roles: Optional[Sequence[str]] = None,
required_actions: Optional[Sequence[str]] = None,
username: Optional[str] = None):
"""
Keycloak User REST object.
:param Mapping[str, Sequence[str]] attributes: A set of Attributes.
:param Mapping[str, Sequence[str]] client_roles: A set of Client Roles.
:param Sequence['KeycloakUserSpecUserCredentialsArgs'] credentials: A set of Credentials.
:param str email: Email.
:param bool email_verified: True if email has already been verified.
:param bool enabled: User enabled flag.
:param Sequence['KeycloakUserSpecUserFederatedIdentitiesArgs'] federated_identities: A set of Federated Identities.
:param str first_name: First Name.
:param Sequence[str] groups: A set of Groups.
:param str id: User ID.
:param str last_name: Last Name.
:param Sequence[str] realm_roles: A set of Realm Roles.
:param Sequence[str] required_actions: A set of Required Actions.
:param str username: User Name.
"""
if attributes is not None:
pulumi.set(__self__, "attributes", attributes)
if client_roles is not None:
pulumi.set(__self__, "client_roles", client_roles)
if credentials is not None:
pulumi.set(__self__, "credentials", credentials)
if email is not None:
pulumi.set(__self__, "email", email)
if email_verified is not None:
pulumi.set(__self__, "email_verified", email_verified)
if enabled is not None:
pulumi.set(__self__, "enabled", enabled)
if federated_identities is not None:
pulumi.set(__self__, "federated_identities", federated_identities)
if first_name is not None:
pulumi.set(__self__, "first_name", first_name)
if groups is not None:
pulumi.set(__self__, "groups", groups)
if id is not None:
pulumi.set(__self__, "id", id)
if last_name is not None:
pulumi.set(__self__, "last_name", last_name)
if realm_roles is not None:
pulumi.set(__self__, "realm_roles", realm_roles)
if required_actions is not None:
pulumi.set(__self__, "required_actions", required_actions)
if username is not None:
pulumi.set(__self__, "username", username)
@property
@pulumi.getter
def attributes(self) -> Optional[Mapping[str, Sequence[str]]]:
"""
A set of Attributes.
"""
return pulumi.get(self, "attributes")
@property
@pulumi.getter(name="clientRoles")
def client_roles(self) -> Optional[Mapping[str, Sequence[str]]]:
"""
A set of Client Roles.
"""
return pulumi.get(self, "client_roles")
@property
@pulumi.getter
def credentials(self) -> Optional[Sequence['outputs.KeycloakUserSpecUserCredentials']]:
"""
A set of Credentials.
"""
return pulumi.get(self, "credentials")
@property
@pulumi.getter
def email(self) -> Optional[str]:
"""
Email.
"""
return pulumi.get(self, "email")
@property
@pulumi.getter(name="emailVerified")
def email_verified(self) -> Optional[bool]:
"""
True if email has already been verified.
"""
return pulumi.get(self, "email_verified")
@property
@pulumi.getter
def enabled(self) -> Optional[bool]:
"""
User enabled flag.
"""
return pulumi.get(self, "enabled")
@property
@pulumi.getter(name="federatedIdentities")
def federated_identities(self) -> Optional[Sequence['outputs.KeycloakUserSpecUserFederatedIdentities']]:
"""
A set of Federated Identities.
"""
return pulumi.get(self, "federated_identities")
@property
@pulumi.getter(name="firstName")
def first_name(self) -> Optional[str]:
"""
First Name.
"""
return pulumi.get(self, "first_name")
@property
@pulumi.getter
def groups(self) -> Optional[Sequence[str]]:
"""
A set of Groups.
"""
return pulumi.get(self, "groups")
@property
@pulumi.getter
def id(self) -> Optional[str]:
"""
User ID.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter(name="lastName")
def last_name(self) -> Optional[str]:
"""
Last Name.
"""
return pulumi.get(self, "last_name")
@property
@pulumi.getter(name="realmRoles")
def realm_roles(self) -> Optional[Sequence[str]]:
"""
A set of Realm Roles.
"""
return pulumi.get(self, "realm_roles")
@property
@pulumi.getter(name="requiredActions")
def required_actions(self) -> Optional[Sequence[str]]:
"""
A set of Required Actions.
"""
return pulumi.get(self, "required_actions")
@property
@pulumi.getter
def username(self) -> Optional[str]:
"""
User Name.
"""
return pulumi.get(self, "username")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakUserSpecUserCredentials(dict):
def __init__(__self__, *,
temporary: Optional[bool] = None,
type: Optional[str] = None,
value: Optional[str] = None):
"""
:param bool temporary: True if this credential object is temporary.
:param str type: Credential Type.
:param str value: Credential Value.
"""
if temporary is not None:
pulumi.set(__self__, "temporary", temporary)
if type is not None:
pulumi.set(__self__, "type", type)
if value is not None:
pulumi.set(__self__, "value", value)
@property
@pulumi.getter
def temporary(self) -> Optional[bool]:
"""
True if this credential object is temporary.
"""
return pulumi.get(self, "temporary")
@property
@pulumi.getter
def type(self) -> Optional[str]:
"""
Credential Type.
"""
return pulumi.get(self, "type")
@property
@pulumi.getter
def value(self) -> Optional[str]:
"""
Credential Value.
"""
return pulumi.get(self, "value")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakUserSpecUserFederatedIdentities(dict):
def __init__(__self__, *,
identity_provider: Optional[str] = None,
user_id: Optional[str] = None,
user_name: Optional[str] = None):
"""
:param str identity_provider: Federated Identity Provider.
:param str user_id: Federated Identity User ID.
:param str user_name: Federated Identity User Name.
"""
if identity_provider is not None:
pulumi.set(__self__, "identity_provider", identity_provider)
if user_id is not None:
pulumi.set(__self__, "user_id", user_id)
if user_name is not None:
pulumi.set(__self__, "user_name", user_name)
@property
@pulumi.getter(name="identityProvider")
def identity_provider(self) -> Optional[str]:
"""
Federated Identity Provider.
"""
return pulumi.get(self, "identity_provider")
@property
@pulumi.getter(name="userId")
def user_id(self) -> Optional[str]:
"""
Federated Identity User ID.
"""
return pulumi.get(self, "user_id")
@property
@pulumi.getter(name="userName")
def user_name(self) -> Optional[str]:
"""
Federated Identity User Name.
"""
return pulumi.get(self, "user_name")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
@pulumi.output_type
class KeycloakUserStatus(dict):
"""
KeycloakUserStatus defines the observed state of KeycloakUser.
"""
def __init__(__self__, *,
message: str,
phase: str):
"""
KeycloakUserStatus defines the observed state of KeycloakUser.
:param str message: Human-readable message indicating details about current operator phase or error.
:param str phase: Current phase of the operator.
"""
pulumi.set(__self__, "message", message)
pulumi.set(__self__, "phase", phase)
@property
@pulumi.getter
def message(self) -> str:
"""
Human-readable message indicating details about current operator phase or error.
"""
return pulumi.get(self, "message")
@property
@pulumi.getter
def phase(self) -> str:
"""
Current phase of the operator.
"""
return pulumi.get(self, "phase")
def _translate_property(self, prop):
return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop
| 40.08212 | 638 | 0.645313 | 14,415 | 129,345 | 5.596393 | 0.040444 | 0.024655 | 0.035613 | 0.05205 | 0.857906 | 0.833127 | 0.81581 | 0.788179 | 0.779986 | 0.768916 | 0 | 0.000461 | 0.262329 | 129,345 | 3,226 | 639 | 40.094544 | 0.845033 | 0.325679 | 0 | 0.802602 | 1 | 0 | 0.135838 | 0.06598 | 0 | 0 | 0 | 0.00248 | 0 | 1 | 0.17138 | false | 0 | 0.003394 | 0.023756 | 0.346154 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 |
cb7564450627b8e178cddf0547e6a56a08190805 | 3,473 | py | Python | netbox/ipam/migrations/0056_standardize_id_fields.py | cybarox/netbox | ea197eff5f4fe925bb354d1375912decd81752bd | [
"Apache-2.0"
] | null | null | null | netbox/ipam/migrations/0056_standardize_id_fields.py | cybarox/netbox | ea197eff5f4fe925bb354d1375912decd81752bd | [
"Apache-2.0"
] | null | null | null | netbox/ipam/migrations/0056_standardize_id_fields.py | cybarox/netbox | ea197eff5f4fe925bb354d1375912decd81752bd | [
"Apache-2.0"
] | null | null | null | from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('ipam', '0055_servicetemplate'),
]
operations = [
# Model IDs
migrations.AlterField(
model_name='aggregate',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='asn',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='fhrpgroup',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='fhrpgroupassignment',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='ipaddress',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='iprange',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='prefix',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='rir',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='role',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='routetarget',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='service',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='servicetemplate',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='vlan',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='vlangroup',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
migrations.AlterField(
model_name='vrf',
name='id',
field=models.BigAutoField(auto_created=True, primary_key=True, serialize=False),
),
# GFK IDs
migrations.AlterField(
model_name='fhrpgroupassignment',
name='interface_id',
field=models.PositiveBigIntegerField(),
),
migrations.AlterField(
model_name='ipaddress',
name='assigned_object_id',
field=models.PositiveBigIntegerField(blank=True, null=True),
),
]
| 34.73 | 92 | 0.580478 | 316 | 3,473 | 6.218354 | 0.151899 | 0.173028 | 0.216285 | 0.250891 | 0.836132 | 0.818321 | 0.756234 | 0.756234 | 0.756234 | 0.756234 | 0 | 0.001663 | 0.307515 | 3,473 | 99 | 93 | 35.080808 | 0.815385 | 0.004895 | 0 | 0.73913 | 0 | 0 | 0.066609 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0.021739 | 0.01087 | 0 | 0.043478 | 0 | 0 | 0 | 0 | null | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 9 |
38116738d8045592107d3e42e173ba9ef95f4212 | 19,855 | py | Python | zaqar_tempest_plugin/services/messaging/json/messaging_client.py | openstack/zaqar-tempest-plugin | 3813c99e501a1f67a9cbde751cf71747a61f6786 | [
"Apache-2.0"
] | 5 | 2017-10-31T13:38:16.000Z | 2019-01-28T22:14:03.000Z | zaqar_tempest_plugin/services/messaging/json/messaging_client.py | openstack/zaqar-tempest-plugin | 3813c99e501a1f67a9cbde751cf71747a61f6786 | [
"Apache-2.0"
] | null | null | null | zaqar_tempest_plugin/services/messaging/json/messaging_client.py | openstack/zaqar-tempest-plugin | 3813c99e501a1f67a9cbde751cf71747a61f6786 | [
"Apache-2.0"
] | null | null | null | # Copyright (c) 2014 Rackspace, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import urllib
from oslo_utils import uuidutils
from oslo_serialization import jsonutils as json
from tempest.lib.common import rest_client
from zaqar_tempest_plugin.api_schema.response.v1 \
import queues as v1schema
from zaqar_tempest_plugin.api_schema.response.v1_1 \
import queues as v11schema
from zaqar_tempest_plugin.api_schema.response.v2 \
import queues as v2schema
class MessagingClient(rest_client.RestClient):
def __init__(self, auth_provider, service, region, **kwargs):
super(MessagingClient, self).__init__(
auth_provider, service, region, **kwargs)
self.version = '1'
self.uri_prefix = 'v{0}'.format(self.version)
client_id = uuidutils.generate_uuid(dashed=False)
self.headers = {'Client-ID': client_id}
class V1MessagingClient(MessagingClient):
def __init__(self, auth_provider, service, region, **kwargs):
super(V1MessagingClient, self).__init__(
auth_provider, service, region, **kwargs)
self.version = '1'
def list_queues(self):
uri = '{0}/queues'.format(self.uri_prefix)
resp, body = self.get(uri)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v1schema.list_queues, resp, body)
return resp, body
def create_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.put(uri, body=None)
self.expected_success(201, resp.status)
return resp, body
def show_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri)
self.expected_success(204, resp.status)
return resp, body
def head_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.head(uri)
self.expected_success(204, resp.status)
return resp, body
def delete_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.delete(uri)
self.expected_success(204, resp.status)
return resp, body
def show_queue_stats(self, queue_name):
uri = '{0}/queues/{1}/stats'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri)
body = json.loads(body)
self.validate_response(v1schema.queue_stats, resp, body)
return resp, body
def show_queue_metadata(self, queue_name):
uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri)
self.expected_success(200, resp.status)
body = json.loads(body)
return resp, body
def set_queue_metadata(self, queue_name, rbody):
uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name)
resp, body = self.put(uri, body=json.dumps(rbody))
self.expected_success(204, resp.status)
return resp, body
def post_messages(self, queue_name, rbody):
uri = '{0}/queues/{1}/messages'.format(self.uri_prefix, queue_name)
resp, body = self.post(uri, body=json.dumps(rbody),
extra_headers=True,
headers=self.headers)
body = json.loads(body)
self.validate_response(v1schema.post_messages, resp, body)
return resp, body
def list_messages(self, queue_name):
uri = '{0}/queues/{1}/messages?echo=True'.format(self.uri_prefix,
queue_name)
resp, body = self.get(uri, extra_headers=True, headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v1schema.list_messages, resp, body)
return resp, body
def show_single_message(self, message_uri):
resp, body = self.get(message_uri, extra_headers=True,
headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v1schema.get_single_message, resp,
body)
return resp, body
def show_multiple_messages(self, message_uri):
resp, body = self.get(message_uri, extra_headers=True,
headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v1schema.get_multiple_messages,
resp,
body)
return resp, body
def delete_messages(self, message_uri):
resp, body = self.delete(message_uri)
self.expected_success(204, resp.status)
return resp, body
def post_claims(self, queue_name, rbody, url_params=False):
uri = '{0}/queues/{1}/claims'.format(self.uri_prefix, queue_name)
if url_params:
uri += '?%s' % urllib.parse.urlencode(url_params)
resp, body = self.post(uri, body=json.dumps(rbody),
extra_headers=True,
headers=self.headers)
body = json.loads(body)
self.validate_response(v1schema.claim_messages, resp, body)
return resp, body
def query_claim(self, claim_uri):
resp, body = self.get(claim_uri)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v1schema.query_claim, resp, body)
return resp, body
def update_claim(self, claim_uri, rbody):
resp, body = self.patch(claim_uri, body=json.dumps(rbody))
self.expected_success(204, resp.status)
return resp, body
def delete_claim(self, claim_uri):
resp, body = self.delete(claim_uri)
self.expected_success(204, resp.status)
return resp, body
class V11MessagingClient(MessagingClient):
def __init__(self, auth_provider, service, region, **kwargs):
super(V11MessagingClient, self).__init__(
auth_provider, service, region, **kwargs)
self.version = '1.1'
self.uri_prefix = 'v{0}'.format(self.version)
def list_queues(self):
uri = '{0}/queues'.format(self.uri_prefix)
resp, body = self.get(uri, headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v11schema.list_queues, resp, body)
return resp, body
def create_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.put(uri, body=None, headers=self.headers)
self.expected_success(201, resp.status)
return resp, body
def show_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri, headers=self.headers)
self.expected_success(200, resp.status)
return resp, body
def delete_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.delete(uri, headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def show_queue_stats(self, queue_name):
uri = '{0}/queues/{1}/stats'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri, headers=self.headers)
body = json.loads(body)
self.validate_response(v11schema.queue_stats, resp, body)
return resp, body
def show_queue_metadata(self, queue_name):
uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri, headers=self.headers)
self.expected_success(200, resp.status)
body = json.loads(body)
return resp, body
def set_queue_metadata(self, queue_name, rbody):
uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name)
resp, body = self.put(uri, body=json.dumps(rbody),
headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def post_messages(self, queue_name, rbody):
uri = '{0}/queues/{1}/messages'.format(self.uri_prefix, queue_name)
resp, body = self.post(uri, body=json.dumps(rbody),
extra_headers=True,
headers=self.headers)
body = json.loads(body)
self.validate_response(v11schema.post_messages, resp, body)
return resp, body
def list_messages(self, queue_name):
uri = '{0}/queues/{1}/messages?echo=True'.format(self.uri_prefix,
queue_name)
resp, body = self.get(uri, extra_headers=True, headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v11schema.list_messages, resp, body)
return resp, body
def show_single_message(self, message_uri):
resp, body = self.get(message_uri, extra_headers=True,
headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v11schema.get_single_message, resp,
body)
return resp, body
def show_multiple_messages(self, message_uri):
resp, body = self.get(message_uri, extra_headers=True,
headers=self.headers)
if resp['status'] != '404':
body = json.loads(body)
self.validate_response(v11schema.get_multiple_messages,
resp,
body)
return resp, body
def delete_messages(self, message_uri):
resp, body = self.delete(message_uri, headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def post_claims(self, queue_name, rbody, url_params=False):
uri = '{0}/queues/{1}/claims'.format(self.uri_prefix, queue_name)
if url_params:
uri += '?%s' % urllib.parse.urlencode(url_params)
resp, body = self.post(uri, body=json.dumps(rbody),
extra_headers=True,
headers=self.headers)
body = json.loads(body)
self.validate_response(v11schema.claim_messages, resp, body)
return resp, body
def query_claim(self, claim_uri):
resp, body = self.get(claim_uri, headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v11schema.query_claim, resp, body)
return resp, body
def update_claim(self, claim_uri, rbody):
resp, body = self.patch(claim_uri, body=json.dumps(rbody),
headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def delete_claim(self, claim_uri):
resp, body = self.delete(claim_uri, headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
class V2MessagingClient(MessagingClient):
def __init__(self, auth_provider, service, region, **kwargs):
super(V2MessagingClient, self).__init__(
auth_provider, service, region, **kwargs)
self.version = '2'
self.uri_prefix = 'v{0}'.format(self.version)
def list_queues(self, url_params=False):
uri = '{0}/queues'.format(self.uri_prefix)
if url_params:
uri += '?%s' % urllib.parse.urlencode(url_params)
resp, body = self.get(uri, headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v2schema.list_queues, resp, body)
return resp, body
def create_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.put(uri, body=None, headers=self.headers)
self.expected_success(201, resp.status)
return resp, body
def show_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri, headers=self.headers)
self.expected_success(200, resp.status)
return resp, body
def delete_queue(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.delete(uri, headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def purge_queue(self, queue_name, resource=None):
uri = '{0}/queues/{1}/purge'.format(self.uri_prefix, queue_name)
rbody = {"resource_types": ["messages", "subscriptions"]}
if resource:
rbody = {"resource_types": resource}
resp, body = self.post(uri, body=json.dumps(rbody),
headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def show_queue_stats(self, queue_name):
uri = '{0}/queues/{1}/stats'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri, headers=self.headers)
body = json.loads(body)
self.validate_response(v2schema.queue_stats, resp, body)
return resp, body
def show_queue_metadata(self, queue_name):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
resp, body = self.get(uri, headers=self.headers)
self.expected_success(200, resp.status)
body = json.loads(body)
return resp, body
def set_queue_metadata(self, queue_name, rbody):
uri = '{0}/queues/{1}'.format(self.uri_prefix, queue_name)
headers = self.headers.copy()
headers['Content-Type'] =\
'application/openstack-messaging-v2.0-json-patch'
resp, body = self.patch(uri, body=json.dumps(rbody),
headers=headers)
self.expected_success(200, resp.status)
return resp, body
def post_messages(self, queue_name, rbody):
uri = '{0}/queues/{1}/messages'.format(self.uri_prefix, queue_name)
resp, body = self.post(uri, body=json.dumps(rbody),
extra_headers=True,
headers=self.headers)
body = json.loads(body)
self.validate_response(v2schema.post_messages, resp, body)
return resp, body
def list_messages(self, queue_name):
uri = '{0}/queues/{1}/messages?echo=True'.format(self.uri_prefix,
queue_name)
resp, body = self.get(uri, extra_headers=True, headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v2schema.list_messages, resp, body)
return resp, body
def show_single_message(self, message_uri):
resp, body = self.get(message_uri, extra_headers=True,
headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v2schema.get_single_message, resp,
body)
return resp, body
def show_multiple_messages(self, message_uri):
resp, body = self.get(message_uri, extra_headers=True,
headers=self.headers)
if resp['status'] != '404':
body = json.loads(body)
self.validate_response(v2schema.get_multiple_messages,
resp,
body)
return resp, body
def delete_messages(self, message_uri):
resp, body = self.delete(message_uri, headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def post_claims(self, queue_name, rbody, url_params=False):
uri = '{0}/queues/{1}/claims'.format(self.uri_prefix, queue_name)
if url_params:
uri += '?%s' % urllib.parse.urlencode(url_params)
resp, body = self.post(uri, body=json.dumps(rbody),
extra_headers=True,
headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v2schema.claim_messages, resp, body)
return resp, body
def query_claim(self, claim_uri):
resp, body = self.get(claim_uri, headers=self.headers)
if resp['status'] != '204':
body = json.loads(body)
self.validate_response(v2schema.query_claim, resp, body)
return resp, body
def update_claim(self, claim_uri, rbody):
resp, body = self.patch(claim_uri, body=json.dumps(rbody),
headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def delete_claim(self, claim_uri):
resp, body = self.delete(claim_uri, headers=self.headers)
self.expected_success(204, resp.status)
return resp, body
def create_subscription(self, queue_name, rbody):
uri = '{0}/queues/{1}/subscriptions'.format(self.uri_prefix,
queue_name)
resp, body = self.post(uri, body=json.dumps(rbody),
extra_headers=True,
headers=self.headers)
body = json.loads(body)
self.validate_response(v2schema.create_subscription, resp, body)
return resp, body
def delete_subscription(self, queue_name, subscription_id):
uri = '{0}/queues/{1}/subscriptions/{2}'.format(self.uri_prefix,
queue_name,
subscription_id)
resp, body = self.delete(uri, headers=self.headers)
return resp, body
def list_subscription(self, queue_name):
uri = '{0}/queues/{1}/subscriptions'.format(self.uri_prefix,
queue_name)
resp, body = self.get(uri, headers=self.headers)
body = json.loads(body)
self.validate_response(v2schema.list_subscriptions, resp, body)
return resp, body
def show_subscription(self, queue_name, subscription_id):
uri = '{0}/queues/{1}/subscriptions/{2}'.format(self.uri_prefix,
queue_name,
subscription_id)
resp, body = self.get(uri, headers=self.headers)
body = json.loads(body)
self.validate_response(v2schema.show_single_subscription, resp, body)
return resp, body
def update_subscription(self, queue_name, subscription_id, rbody):
uri = '{0}/queues/{1}/subscriptions/{2}'.format(self.uri_prefix,
queue_name,
subscription_id)
resp, body = self.patch(uri, body=json.dumps(rbody),
headers=self.headers)
return resp, body
| 38.779297 | 77 | 0.59592 | 2,380 | 19,855 | 4.805882 | 0.071849 | 0.095821 | 0.057702 | 0.077286 | 0.901207 | 0.90007 | 0.890103 | 0.872967 | 0.856968 | 0.842368 | 0 | 0.019151 | 0.292571 | 19,855 | 511 | 78 | 38.855186 | 0.795173 | 0.027953 | 0 | 0.813131 | 0 | 0 | 0.054135 | 0.027068 | 0 | 0 | 0 | 0 | 0 | 1 | 0.14899 | false | 0 | 0.017677 | 0 | 0.315657 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
3836019055aa4cd174d05be7e971ed828f0eed54 | 12,757 | py | Python | CIFAR10/more_attack.py | haiphanNJIT/StoBatch | 08a6505b1fad609850de6c6bbaaa54c0200f0adb | [
"MIT"
] | 6 | 2020-06-13T17:05:52.000Z | 2021-12-02T12:41:39.000Z | MNIST/more_attack.py | haiphanNJIT/StoBatch | 08a6505b1fad609850de6c6bbaaa54c0200f0adb | [
"MIT"
] | null | null | null | MNIST/more_attack.py | haiphanNJIT/StoBatch | 08a6505b1fad609850de6c6bbaaa54c0200f0adb | [
"MIT"
] | 1 | 2021-04-05T11:20:22.000Z | 2021-04-05T11:20:22.000Z | import numpy as np
from six.moves import xrange
import tensorflow as tf
from cleverhans.attacks_tf import fgm, fgsm
from build_utils import batch_adv
def model_loss(y, model, mean=True):
"""
FROM cleverhans/utils_tf
Define loss of TF graph
:param y: correct labels
:param model: output of the model
:param mean: boolean indicating whether should return mean of loss
or vector of losses for each input of the batch
:return: return mean of loss if True, otherwise return vector with per
sample loss
"""
op = model.op
if "softmax" in str(op).lower():
logits, = op.inputs
else:
logits = model
out = tf.nn.softmax_cross_entropy_with_logits(logits=logits, labels=y)
if mean:
out = tf.reduce_mean(out)
return out
def fgm_pre_computed_grad(x, grad, eps=0.3, ord=np.inf,
clip_min=None, clip_max=None,
targeted=False):
"""
TensorFlow implementation of the Fast Gradient Method using pre computed gradients.
:param x: input
:param grad: pre-computed gradients for x on the pre-trained model
(use negative (flipped) loss for the gradient if targeted)
:param eps: the epsilon (input variation parameter)
:param ord: (optional) Order of the norm (mimics NumPy).
Possible values: np.inf, 1 or 2.
:param clip_min: Minimum float value for adversarial example components
:param clip_max: Maximum float value for adversarial example components
:param targeted: Is the attack targeted or untargeted? Untargeted, the
default, will try to make the label incorrect. Targeted
will instead try to move in the direction of being more
like y.
:return: a tensor for the adversarial example
"""
if ord == np.inf:
# Take sign of gradient
normalized_grad = tf.sign(grad)
# The following line should not change the numerical results.
# It applies only because `normalized_grad` is the output of
# a `sign` op, which has zero derivative anyway.
# It should not be applied for the other norms, where the
# perturbation has a non-zero derivative.
normalized_grad = tf.stop_gradient(normalized_grad)
elif ord == 1:
red_ind = list(xrange(1, len(x.get_shape())))
normalized_grad = grad / tf.reduce_sum(tf.abs(grad),
reduction_indices=red_ind,
keep_dims=True)
elif ord == 2:
red_ind = list(xrange(1, len(x.get_shape())))
square = tf.reduce_sum(tf.square(grad),
reduction_indices=red_ind,
keep_dims=True)
normalized_grad = grad / tf.sqrt(square)
else:
raise NotImplementedError("Only L-inf, L1 and L2 norms are "
"currently implemented.")
# Multiply by constant epsilon
scaled_grad = eps * normalized_grad
# Add perturbation to original example to obtain adversarial example
adv_x = x + scaled_grad
# If clipping is needed, reset all values outside of [clip_min, clip_max]
if (clip_min is not None) and (clip_max is not None):
adv_x = tf.clip_by_value(adv_x, clip_min, clip_max)
return adv_x
def rand_fgm(sess, x, logits, y=None, eps=0.3, ord=np.inf, rand_eps=0.3, rand_alpha=0.05,
clip_min=None, clip_max=None,
targeted=False):
"""
TensorFlow implementation of the Fast Gradient Method.
:param x: the input placeholder
:param preds: the model's output tensor (the attack expects the
probabilities, i.e., the output of the softmax)
:param y: (optional) A placeholder for the model labels. If targeted
is true, then provide the target label. Otherwise, only provide
this parameter if you'd like to use true labels when crafting
adversarial samples. Otherwise, model predictions are used as
labels to avoid the "label leaking" effect (explained in this
paper: https://arxiv.org/abs/1611.01236). Default is None.
Labels should be one-hot-encoded.
:param eps: the epsilon (input variation parameter)
:param ord: (optional) Order of the norm (mimics NumPy).
Possible values: np.inf, 1 or 2.
:param clip_min: Minimum float value for adversarial example components
:param clip_max: Maximum float value for adversarial example components
:param targeted: Is the attack targeted or untargeted? Untargeted, the
default, will try to make the label incorrect. Targeted
will instead try to move in the direction of being more
like y.
:return: a tensor for the adversarial example
"""
x_rand = x + rand_alpha * tf.sign(tf.random_normal(shape=tf.get_shape(x), mean=0.0, stddev=1.0))
if y is None:
# Using model predictions as ground truth to avoid label leaking
preds_max = tf.reduce_max(preds, 1, keep_dims=True)
y = tf.to_float(tf.equal(preds, preds_max))
y = tf.stop_gradient(y)
y = y / tf.reduce_sum(y, 1, keep_dims=True)
# Compute loss
loss = utils_tf.model_loss(y, preds, mean=False)
if targeted:
loss = -loss
# Define gradient of loss wrt input
grad, = tf.gradients(loss, x)
if ord == np.inf:
# Take sign of gradient
normalized_grad = tf.sign(grad)
# The following line should not change the numerical results.
# It applies only because `normalized_grad` is the output of
# a `sign` op, which has zero derivative anyway.
# It should not be applied for the other norms, where the
# perturbation has a non-zero derivative.
normalized_grad = tf.stop_gradient(normalized_grad)
elif ord == 1:
red_ind = list(xrange(1, len(x.get_shape())))
normalized_grad = grad / tf.reduce_sum(tf.abs(grad),
reduction_indices=red_ind,
keep_dims=True)
elif ord == 2:
red_ind = list(xrange(1, len(x.get_shape())))
square = tf.reduce_sum(tf.square(grad),
reduction_indices=red_ind,
keep_dims=True)
normalized_grad = grad / tf.sqrt(square)
else:
raise NotImplementedError("Only L-inf, L1 and L2 norms are "
"currently implemented.")
# Multiply by constant epsilon
scaled_grad = eps * normalized_grad
# Add perturbation to original example to obtain adversarial example
adv_x = x + scaled_grad
# If clipping is needed, reset all values outside of [clip_min, clip_max]
if (clip_min is not None) and (clip_max is not None):
adv_x = tf.clip_by_value(adv_x, clip_min, clip_max)
return adv_x
def iter_fgsm(sess, x_input_t, labels_t, x_input, labels, batch_size,
preds_t, target_labels_t,
steps, total_eps, step_eps,
clip_min=0.0, clip_max=1.0,
ord=np.inf, targeted=False):
"""
I-FGSM attack. This function directly generate adv inputs
"""
eta_t = fgm(x_input_t, preds_t, y=target_labels_t, eps=step_eps, ord=ord,
clip_min=clip_min, clip_max=clip_max, targeted=targeted) - x_input_t
if ord == np.inf:
eta_t = tf.clip_by_value(eta_t, -total_eps, total_eps)
elif ord in [1, 2]:
reduc_ind = list(xrange(1, len(tf.shape(eta_t))))
if ord == 1:
norm = tf.reduce_sum(tf.abs(eta_t),
reduction_indices=reduc_ind,
keep_dims=True)
elif ord == 2:
norm = tf.sqrt(tf.reduce_sum(tf.square(eta_t),
reduction_indices=reduc_ind,
keep_dims=True))
eta_t = eta_t * total_eps / norm
x_adv_t = x_input_t + eta_t
x_adv = x_input
for i in range(steps):
x_adv = batch_adv(sess, x_adv_t, x_input_t, labels_t, x_adv, labels, batch_size)
return adv_x
def iter_fgsm_t(x_input_t, preds_t, target_labels_t,
steps, total_eps, step_eps,
clip_min=0.0, clip_max=1.0, ord=np.inf, targeted=False):
"""
I-FGSM attack.
"""
eta_t = fgm(x_input_t, preds_t, y=target_labels_t, eps=step_eps, ord=ord,
clip_min=clip_min, clip_max=clip_max, targeted=targeted) - x_input_t
if ord == np.inf:
eta_t = tf.clip_by_value(eta_t, -total_eps, total_eps)
elif ord in [1, 2]:
reduc_ind = list(xrange(1, len(tf.shape(eta_t))))
if ord == 1:
norm = tf.reduce_sum(tf.abs(eta_t),
reduction_indices=reduc_ind,
keep_dims=True)
elif ord == 2:
norm = tf.sqrt(tf.reduce_sum(tf.square(eta_t),
reduction_indices=reduc_ind,
keep_dims=True))
eta_t = eta_t * total_eps / norm
x_adv_t = x_input_t + eta_t
return x_adv_t
def _fgm(x, preds, y=None, eps=0.3, ord=np.inf,
clip_min=None, clip_max=None,
targeted=False):
"""
TensorFlow implementation of the Fast Gradient Method.
:param x: the input placeholder
:param preds: the model's output tensor (the attack expects the
probabilities, i.e., the output of the softmax)
:param y: (optional) A placeholder for the model labels. If targeted
is true, then provide the target label. Otherwise, only provide
this parameter if you'd like to use true labels when crafting
adversarial samples. Otherwise, model predictions are used as
labels to avoid the "label leaking" effect (explained in this
paper: https://arxiv.org/abs/1611.01236). Default is None.
Labels should be one-hot-encoded.
:param eps: the epsilon (input variation parameter)
:param ord: (optional) Order of the norm (mimics NumPy).
Possible values: np.inf, 1 or 2.
:param clip_min: Minimum float value for adversarial example components
:param clip_max: Maximum float value for adversarial example components
:param targeted: Is the attack targeted or untargeted? Untargeted, the
default, will try to make the label incorrect. Targeted
will instead try to move in the direction of being more
like y.
:return: a tensor for the adversarial example
"""
if y is None:
# Using model predictions as ground truth to avoid label leaking
preds_max = tf.reduce_max(preds, 1, keep_dims=True)
y = tf.to_float(tf.equal(preds, preds_max))
y = tf.stop_gradient(y)
y = y / tf.reduce_sum(y, 1, keep_dims=True)
# Compute loss
loss = utils_tf.model_loss(y, preds, mean=False)
if targeted:
loss = -loss
# Define gradient of loss wrt input
grad, = tf.gradients(loss, x)
if ord == np.inf:
# Take sign of gradient
normalized_grad = tf.sign(grad)
# The following line should not change the numerical results.
# It applies only because `normalized_grad` is the output of
# a `sign` op, which has zero derivative anyway.
# It should not be applied for the other norms, where the
# perturbation has a non-zero derivative.
normalized_grad = tf.stop_gradient(normalized_grad)
elif ord == 1:
red_ind = list(xrange(1, len(x.get_shape())))
normalized_grad = grad / tf.reduce_sum(tf.abs(grad),
reduction_indices=red_ind,
keep_dims=True)
elif ord == 2:
red_ind = list(xrange(1, len(x.get_shape())))
square = tf.reduce_sum(tf.square(grad),
reduction_indices=red_ind,
keep_dims=True)
normalized_grad = grad / tf.sqrt(square)
else:
raise NotImplementedError("Only L-inf, L1 and L2 norms are "
"currently implemented.")
# Multiply by constant epsilon
scaled_grad = eps * normalized_grad
# Add perturbation to original example to obtain adversarial example
adv_x = x + scaled_grad
# If clipping is needed, reset all values outside of [clip_min, clip_max]
if (clip_min is not None) and (clip_max is not None):
adv_x = tf.clip_by_value(adv_x, clip_min, clip_max)
return adv_x
| 42.38206 | 100 | 0.612997 | 1,780 | 12,757 | 4.241011 | 0.134831 | 0.019473 | 0.022255 | 0.017221 | 0.874156 | 0.873228 | 0.86475 | 0.86475 | 0.861704 | 0.861704 | 0 | 0.008961 | 0.308928 | 12,757 | 300 | 101 | 42.523333 | 0.847323 | 0.418907 | 0 | 0.818182 | 0 | 0 | 0.023972 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.038961 | false | 0 | 0.032468 | 0 | 0.11039 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
69944ec5508df035d8cb6588b4a276d892b033bc | 51,749 | py | Python | server/config/mv_fantasy/templates.py | AustralianDisabilityLimited/MultiversePlatform | 7e1aad33d48b9e47f3db2ca638cb57592336ddb7 | [
"MIT"
] | 33 | 2015-02-16T02:52:08.000Z | 2022-02-18T08:46:32.000Z | server/config/mv_fantasy/templates.py | bensku/MultiversePlatform | 7e1aad33d48b9e47f3db2ca638cb57592336ddb7 | [
"MIT"
] | 1 | 2017-09-09T18:50:23.000Z | 2020-12-29T18:13:56.000Z | server/config/mv_fantasy/templates.py | bensku/MultiversePlatform | 7e1aad33d48b9e47f3db2ca638cb57592336ddb7 | [
"MIT"
] | 31 | 2015-02-07T16:20:24.000Z | 2022-02-23T15:02:43.000Z | #
# The Multiverse Platform is made available under the MIT License.
#
# Copyright (c) 2012 The Multiverse Foundation
#
# Permission is hereby granted, free of charge, to any person
# obtaining a copy of this software and associated documentation
# files (the "Software"), to deal in the Software without restriction,
# including without limitation the rights to use, copy, modify,
# merge, publish, distribute, sublicense, and/or sell copies
# of the Software, and to permit persons to whom the Software
# is furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be
# included in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
# OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
# HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE
# OR OTHER DEALINGS IN THE SOFTWARE.
#
#
from multiverse.mars import *
from multiverse.mars.objects import *
from multiverse.mars.core import *
from multiverse.mars.events import *
from multiverse.mars.util import *
from multiverse.mars.plugins import *
from multiverse.server.plugins import *
from multiverse.server.math import *
from multiverse.server.events import *
from multiverse.server.objects import *
from multiverse.server.engine import *
from java.lang import *
True=1
False=0
class TemplateHook(EnginePlugin.PluginActivateHook):
def activate(self):
############################################################
#
# DisplayContexts
#
############################################################
human_female_base_DC = DisplayContext("human_female.mesh", True)
human_female_base_DC.addSubmesh(DisplayContext.Submesh("bodyShape-lib.0",
"human_female.skin_material"))
human_female_base_DC.addSubmesh(DisplayContext.Submesh("head_aShape-lib.0",
"human_female.head_a_material"))
human_female_leather_pantsDC = DisplayContext()
human_female_leather_pantsDC.setMeshFile("human_female.mesh")
human_female_leather_pantsDC.addSubmesh(DisplayContext.Submesh("leather_a_pantsShape-lib.0",
"human_female.leather_a_material"))
human_female_leather_pantsDC.addSubmesh(DisplayContext.Submesh("leather_a_beltShape-lib.0",
"human_female.leather_a_material"))
human_female_leather_tunicDC = DisplayContext()
human_female_leather_tunicDC.setMeshFile("human_female.mesh")
human_female_leather_tunicDC.addSubmesh(DisplayContext.Submesh("leather_a_tunicShape-lib.0",
"human_female.leather_a_material"))
human_female_leather_bootsDC = DisplayContext()
human_female_leather_bootsDC.setMeshFile("human_female.mesh")
human_female_leather_bootsDC.addSubmesh(DisplayContext.Submesh("leather_a_bootsShape-lib.0",
"human_female.leather_a_material"))
human_female_leather_glovesDC = DisplayContext()
human_female_leather_glovesDC.setMeshFile("human_female.mesh")
human_female_leather_glovesDC.addSubmesh(DisplayContext.Submesh("leather_a_bracersShape-lib.0",
"human_female.leather_a_material"))
human_female_plate_tunicDC = DisplayContext()
human_female_plate_tunicDC.setMeshFile("human_female.mesh")
human_female_plate_tunicDC.addSubmesh(DisplayContext.Submesh("plate_b_tunicShape-lib.0",
"human_female.plate_b_material"))
zombie_base_DC = DisplayContext("zombie.mesh")
zombie_base_DC.addSubmesh(DisplayContext.Submesh("Zombie_Body2-obj.0", "Zombie.Zombie_Body"))
zombie_base_DC.addSubmesh(DisplayContext.Submesh("Zombie_Clothes2-obj.0", "Zombie.Zombie_Clothes"))
orc_base_DC = DisplayContext("orc.mesh", True)
human_female_ruth_base_DC = DisplayContext("human_female_ruth.mesh", True)
human_female_ruth_base_DC.addSubmesh(DisplayContext.Submesh("human_female_body_ruth-mesh.0",
"human_female_ruth.ruth_body_clothed_mat"))
human_female_ruth_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth-mesh.0",
"human_female_ruth.ruth_head_mat"))
human_female_ruth_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth_hair-mesh.0",
"human_female_ruth.ruth_hair_mat"))
human_female_ruth_leather_pants_DC = DisplayContext("human_female_ruth.mesh")
human_female_ruth_leather_pants_DC.addSubmesh(DisplayContext.Submesh("leather_armor_legs-mesh.0",
"human_female_ruth.leather_armor_mat"))
human_female_ruth_leather_pants_DC.addSubmesh(DisplayContext.Submesh("leather_armor_belt-mesh.0",
"human_female_ruth.leather_armor_mat"))
human_female_ruth_leather_tunic_DC = DisplayContext("human_female_ruth.mesh")
human_female_ruth_leather_tunic_DC.addSubmesh(DisplayContext.Submesh("leather_armor_chest-mesh.0",
"human_female_ruth.leather_armor_mat"))
human_female_ruth_leather_boots_DC = DisplayContext("human_female_ruth.mesh")
human_female_ruth_leather_boots_DC.addSubmesh(DisplayContext.Submesh("leather_armor_boots-mesh.0",
"human_female_ruth.leather_armor_mat"))
human_female_ruth_leather_gloves_DC = DisplayContext("human_female_ruth.mesh")
human_female_ruth_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_rt-mesh.0",
"human_female_ruth.leather_armor_mat"))
human_female_ruth_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_lt-mesh.0",
"human_female_ruth.leather_armor_mat"))
human_male_DC = DisplayContext("human_male.mesh", True)
human_male_DC.addSubmesh(DisplayContext.Submesh("head_bShape-lib.0",
"human_male.head_a_material"))
human_male_DC.addSubmesh(DisplayContext.Submesh("bodyShape-lib.0",
"human_male.skin_material"))
human_male_DC.addSubmesh(DisplayContext.Submesh("cloth_a_pantsShape-lib.0",
"human_male.cloth_a_material"))
human_male_DC.addSubmesh(DisplayContext.Submesh("cloth_a_bootsShape-lib.0",
"human_male.cloth_a_material"))
human_male_DC.addSubmesh(DisplayContext.Submesh("cloth_a_shirtShape-lib.0",
"human_male.cloth_a_material"))
female_player_01_base_DC = DisplayContext("human_female_fantasy.mesh", True)
female_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_female_body-mesh.0",
"human_female_fantasy.body_01_clothed_mat"))
female_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_01-mesh.0",
"human_female_fantasy.head_01_mat"))
female_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_01_hair-mesh.0",
"human_female_fantasy.head_01_hair_01_mat"))
female_player_02_base_DC = DisplayContext("human_female_fantasy.mesh", True)
female_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_female_body-mesh.0",
"human_female_fantasy.body_02_clothed_mat"))
female_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_02-mesh.0",
"human_female_fantasy.head_02_mat"))
female_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_female_head_02_hair_01-mesh.0",
"human_female_fantasy.head_02_hair_01_mat"))
female_player_leather_pants_DC = DisplayContext("human_female_fantasy.mesh")
female_player_leather_pants_DC.addSubmesh(DisplayContext.Submesh("leather_armor_legs-mesh.0",
"human_female_fantasy.leather_armor_mat"))
female_player_leather_pants_DC.addSubmesh(DisplayContext.Submesh("leather_armor_belt-mesh.0",
"human_female_fantasy.leather_armor_mat"))
female_player_leather_tunic_DC = DisplayContext("human_female_fantasy.mesh")
female_player_leather_tunic_DC.addSubmesh(DisplayContext.Submesh("leather_armor_chest-mesh.0",
"human_female_fantasy.leather_armor_mat"))
female_player_leather_tunic_DC.addSubmesh(DisplayContext.Submesh("leather_armor_jewels-mesh.0",
"human_female_fantasy.leather_armor_mat"))
female_player_leather_boots_DC = DisplayContext("human_female_fantasy.mesh")
female_player_leather_boots_DC.addSubmesh(DisplayContext.Submesh("leather_armor_boot_lt-mesh.0",
"human_female_fantasy.leather_armor_mat"))
female_player_leather_boots_DC.addSubmesh(DisplayContext.Submesh("leather_armor_boot_rt-mesh.0",
"human_female_fantasy.leather_armor_mat"))
female_player_leather_gloves_DC = DisplayContext("human_female_fantasy.mesh")
female_player_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_rt-mesh.0",
"human_female_fantasy.leather_armor_mat"))
female_player_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_lt-mesh.0",
"human_female_fantasy.leather_armor_mat"))
male_player_01_base_DC = DisplayContext("human_male_fantasy.mesh", True)
male_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_male_body-mesh.0",
"human_male_fantasy.human_male_body_01"))
male_player_01_base_DC.addSubmesh(DisplayContext.Submesh("human_male_head_01-mesh.0",
"human_male_fantasy.human_male_head_01"))
male_player_01_base_DC.addSubmesh(DisplayContext.Submesh("male_head_01_hair_01-mesh.0",
"human_male_fantasy.human_male_head_01_hair_01"))
male_player_02_base_DC = DisplayContext("human_male_fantasy.mesh", True)
male_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_male_body-mesh.0",
"human_male_fantasy.human_male_body_02"))
male_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_male_head_02-mesh.0",
"human_male_fantasy.human_male_head_02"))
male_player_02_base_DC.addSubmesh(DisplayContext.Submesh("human_male_02_hair_01-mesh.0",
"human_male_fantasy.human_male_head_02_hair_01"))
male_player_leather_pants_DC = DisplayContext("human_male_fantasy.mesh")
male_player_leather_pants_DC.addSubmesh(DisplayContext.Submesh("male_leather_b_legs-mesh.0",
"human_male_fantasy.human_male_armor_leather_b"))
male_player_leather_tunic_DC = DisplayContext("human_male_fantasy.mesh")
male_player_leather_tunic_DC.addSubmesh(DisplayContext.Submesh("male_leather_b_chest-mesh.0",
"human_male_fantasy.human_male_armor_leather_b"))
male_player_leather_boots_DC = DisplayContext("human_male_fantasy.mesh")
male_player_leather_boots_DC.addSubmesh(DisplayContext.Submesh("male_leather_b_feet-mesh.0",
"human_male_fantasy.human_male_armor_leather_b"))
male_player_leather_gloves_DC = DisplayContext("human_male_fantasy.mesh")
male_player_leather_gloves_DC.addSubmesh(DisplayContext.Submesh("male_leather_b_hands-mesh.0",
"human_male_fantasy.human_male_armor_leather_b"))
hilldale_scout_DC = DisplayContext("human_female_ruth.mesh", True)
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("human_female_body_ruth-mesh.0",
"human_female_ruth.ruth_body_clothed_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth-mesh.0",
"human_female_ruth.ruth_head_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth_hair-mesh.0",
"human_female_ruth.ruth_hair_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_legs-mesh.0",
"human_female_ruth.leather_armor_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_belt-mesh.0",
"human_female_ruth.leather_armor_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_chest-mesh.0",
"human_female_ruth.leather_armor_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_jewel-mesh.0",
"human_female_ruth.leather_armor_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_boots-mesh.0",
"human_female_ruth.leather_armor_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_rt-mesh.0",
"human_female_ruth.leather_armor_mat"))
hilldale_scout_DC.addSubmesh(DisplayContext.Submesh("leather_armor_bracer_lt-mesh.0",
"human_female_ruth.leather_armor_mat"))
constable_dillon_DC = DisplayContext("human_female_ruth.mesh", True)
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("human_female_body_ruth-mesh.0",
"human_female_ruth.ruth_body_clothed_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("human_female_head_ruth-mesh.0",
"human_female_ruth.ruth_head_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxChest-mesh.0",
"human_female_ruth.brax_armor_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxHelmetArmor-mesh.0",
"human_female_ruth.brax_armor_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxLegs-mesh.0",
"human_female_ruth.brax_armor_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxBracerLt-mesh.0",
"human_female_ruth.brax_armor_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxBracerRt-mesh.0",
"human_female_ruth.brax_armor_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxShoulderLt-mesh.0",
"human_female_ruth.brax_armor_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxShoulderRt-mesh.0",
"human_female_ruth.brax_armor_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxBelt-mesh.0",
"human_female_ruth.brax_armor_mat"))
constable_dillon_DC.addSubmesh(DisplayContext.Submesh("braxBoots-mesh.0",
"human_female_ruth.brax_armor_mat"))
brax_DC = DisplayContext("brax.mesh", True)
brax_DC.addSubmesh(DisplayContext.Submesh("braxHeadShape.0", "brax.braxBodyMat"))
brax_DC.addSubmesh(DisplayContext.Submesh("braxBodyShape.0", "brax.braxBodyMat"))
brax_DC.addSubmesh(DisplayContext.Submesh("braxTailShape.0", "brax.braxBodyMat"))
brax_DC.addSubmesh(DisplayContext.Submesh("brax_shell1Shape.0", "brax.braxBodyMat"))
brax_DC.addSubmesh(DisplayContext.Submesh("brax_shell2Shape.0", "brax.braxBodyMat"))
#############################################################
#
# mob templates
#
#############################################################
defaultSlots = MarsEquipInfo("default")
defaultSlots.addEquipSlot(MarsEquipSlot.PRIMARYWEAPON)
#
# Wolf Template
#
tmpl = Template("Wolf")
tmpl.put(WorldManagerClient.NAMESPACE,
WorldManagerClient.TEMPL_DISPLAY_CONTEXT,
DisplayContext("wolf.mesh", True))
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(InventoryClient.NAMESPACE, InventoryClient.TEMPL_ITEMS, "Wolf Skin")
tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18))
tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18))
tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10))
tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10))
tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100))
tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100))
tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0))
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability")
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect")
tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10)
tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5)
ObjectManagerClient.registerTemplate(tmpl)
#
# Coyote Template
#
tmpl = Template("Coyote")
tmpl.put(WorldManagerClient.NAMESPACE,
WorldManagerClient.TEMPL_DISPLAY_CONTEXT,
DisplayContext("wolf.mesh", True))
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_SCALE, MVVector(0.75, 0.75, 0.75))
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18))
tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18))
tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10))
tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10))
tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100))
tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100))
tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0))
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability")
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect")
tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10)
tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5)
ObjectManagerClient.registerTemplate(tmpl)
#
# Crocodile Template
#
tmpl = Template("Crocodile")
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT,
DisplayContext("crocodile.mesh", True))
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_RUN_THRESHOLD, Float(7000))
tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18))
tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18))
tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10))
tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10))
tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100))
tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100))
tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0))
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability")
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect")
tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10)
tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5)
ObjectManagerClient.registerTemplate(tmpl)
#
# Zombie Template
#
tmpl = Template("Zombie")
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, zombie_base_DC)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True))
tmpl.put(InventoryClient.NAMESPACE, InventoryClient.TEMPL_ITEMS, "Zombie Dust")
tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18))
tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18))
tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10))
tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10))
tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100))
tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100))
tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0))
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability")
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect")
tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10)
tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5)
ObjectManagerClient.registerTemplate(tmpl)
#
# Brax Template
#
tmpl = Template("Brax")
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, brax_DC)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_RUN_THRESHOLD, Float(7000))
tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18))
tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18))
tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10))
tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10))
tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100))
tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100))
tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0))
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability")
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect")
tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10)
tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5)
ObjectManagerClient.registerTemplate(tmpl)
#
# Orc Warrior Template
#
tmpl = Template("Orc Warrior")
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, orc_base_DC)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(InventoryClient.NAMESPACE, InventoryClient.TEMPL_ITEMS, "*sword10")
tmpl.put(CombatClient.NAMESPACE, "attackable", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "combat.mobflag", Boolean(True))
tmpl.put(CombatClient.NAMESPACE, "strength", MarsStat("strength", 18))
tmpl.put(CombatClient.NAMESPACE, "agility", MarsStat("agility", 18))
tmpl.put(CombatClient.NAMESPACE, "stamina", MarsStat("stamina", 10))
tmpl.put(CombatClient.NAMESPACE, "intelligence", MarsStat("intelligence", 10))
tmpl.put(CombatClient.NAMESPACE, "health", MarsStat("health", 100))
tmpl.put(CombatClient.NAMESPACE, "mana", MarsStat("mana", 100))
tmpl.put(CombatClient.NAMESPACE, "offense skill", MarsStat("offense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "defense skill", MarsStat("defense skill", 5))
tmpl.put(CombatClient.NAMESPACE, "armor", MarsStat("armor", 0))
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_AUTOATTACK_ABILITY, "attack ability")
tmpl.put(CombatClient.NAMESPACE, CombatInfo.COMBAT_PROP_REGEN_EFFECT, "regen effect")
tmpl.put(CombatClient.NAMESPACE, "weaponBaseDmg", 10)
tmpl.put(CombatClient.NAMESPACE, "weaponVarDmg", 5)
ObjectManagerClient.registerTemplate(tmpl)
#
# Hilldale Scout
#
tmpl = Template("Hilldale Scout")
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, hilldale_scout_DC)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(WorldManagerClient.NAMESPACE,
WorldManagerClient.TEMPL_ORIENT,
Quaternion(0, 0.468, 0, 0.884))
tmpl.put(InventoryClient.NAMESPACE, InventoryClient.TEMPL_ITEMS,
"*Leather Tunic; *Leather Pants; *Leather Boots; *sword4")
ObjectManagerClient.registerTemplate(tmpl)
#
# Constable Dillon
#
tmpl = Template("Constable Dillon")
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, constable_dillon_DC)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(WorldManagerClient.NAMESPACE,
WorldManagerClient.TEMPL_ORIENT,
Quaternion(0, 0.857, 0, -0.515))
ObjectManagerClient.registerTemplate(tmpl)
#
# Cyrus Blackfire
#
tmpl = Template("Cyrus Blackfire")
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_DISPLAY_CONTEXT, human_male_DC)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_NAME, "Cyrus Blackfire")
tmpl.put(WorldManagerClient.NAMESPACE,
WorldManagerClient.TEMPL_ORIENT,
Quaternion(0, 0.857, 0, -0.515))
ObjectManagerClient.registerTemplate(tmpl)
#
# Weapon Templates
#
equipInfo = MarsEquipInfo("weapon")
equipInfo.addEquipSlot(MarsEquipSlot.PRIMARYWEAPON)
dc = DisplayContext("sword.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("Bronze Longsword")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_basic.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword1")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_broad.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword2")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_katar.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword3")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_katareen.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword4")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_leaf.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword5")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_pointy.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword6")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_sabre.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword7")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_serpent.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword8")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_short.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword9")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
dc = DisplayContext("FW_sword_human_stone.mesh")
dc.setAttachInfo(DisplayState.IN_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dc.setAttachInfo(DisplayState.NON_COMBAT,
MarsEquipSlot.PRIMARYWEAPON,
MarsAttachSocket.PRIMARYWEAPON)
dcMap = DCMap()
dcMap.add(human_female_base_DC, dc)
dcMap.add(orc_base_DC, dc)
dcMap.add(zombie_base_DC, dc)
dcMap.add(human_female_ruth_base_DC, dc)
dcMap.add(female_player_01_base_DC, dc)
dcMap.add(female_player_02_base_DC, dc)
dcMap.add(male_player_01_base_DC, dc)
dcMap.add(male_player_02_base_DC, dc)
tmpl = Template("sword10")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\WEAPON_sword_A")
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
tmpl.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(tmpl)
#
# Armor Templates
#
equipInfo = MarsEquipInfo("armor")
equipInfo.addEquipSlot(MarsEquipSlot.CHEST)
dcMap = DCMap()
dcMap.add(human_female_base_DC, human_female_leather_tunicDC)
dcMap.add(human_female_ruth_base_DC, human_female_ruth_leather_tunic_DC)
dcMap.add(female_player_01_base_DC, female_player_leather_tunic_DC)
dcMap.add(female_player_02_base_DC, female_player_leather_tunic_DC)
dcMap.add(male_player_01_base_DC, male_player_leather_tunic_DC)
dcMap.add(male_player_02_base_DC, male_player_leather_tunic_DC)
item = Template("Leather Tunic")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ARMOR_leather_A_chest")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(item)
equipInfo = MarsEquipInfo("pants")
equipInfo.addEquipSlot(MarsEquipSlot.LEGS)
dcMap = DCMap()
dcMap.add(human_female_base_DC, human_female_leather_pantsDC)
dcMap.add(human_female_ruth_base_DC, human_female_ruth_leather_pants_DC)
dcMap.add(female_player_01_base_DC, female_player_leather_pants_DC)
dcMap.add(female_player_02_base_DC, female_player_leather_pants_DC)
dcMap.add(male_player_01_base_DC, male_player_leather_pants_DC)
dcMap.add(male_player_02_base_DC, male_player_leather_pants_DC)
item = Template("Leather Pants")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ARMOR_leather_A_legs")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(item)
equipInfo = MarsEquipInfo("boots")
equipInfo.addEquipSlot(MarsEquipSlot.FEET)
dcMap = DCMap()
dcMap.add(human_female_base_DC, human_female_leather_bootsDC)
dcMap.add(human_female_ruth_base_DC, human_female_ruth_leather_boots_DC)
dcMap.add(female_player_01_base_DC, female_player_leather_boots_DC)
dcMap.add(female_player_02_base_DC, female_player_leather_boots_DC)
dcMap.add(male_player_01_base_DC, male_player_leather_boots_DC)
dcMap.add(male_player_02_base_DC, male_player_leather_boots_DC)
item = Template("Leather Boots")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ARMOR_leather_A_feet")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(item)
equipInfo = MarsEquipInfo("gloves")
equipInfo.addEquipSlot(MarsEquipSlot.HANDS)
dcMap = DCMap()
dcMap.add(human_female_base_DC, human_female_leather_glovesDC)
dcMap.add(human_female_ruth_base_DC, human_female_ruth_leather_gloves_DC)
dcMap.add(female_player_01_base_DC, female_player_leather_gloves_DC)
dcMap.add(female_player_02_base_DC, female_player_leather_gloves_DC)
dcMap.add(male_player_01_base_DC, male_player_leather_gloves_DC)
dcMap.add(male_player_02_base_DC, male_player_leather_gloves_DC)
item = Template("Leather Gloves")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ARMOR_leather_A_hands")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK, EquipActivateHook())
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_EQUIP_INFO, equipInfo)
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_DCMAP, dcMap)
ObjectManagerClient.registerTemplate(item)
#
# Item Templates
#
item = Template("Wolf Skin")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\Icons\INV_hide_basic-pelt")
ObjectManagerClient.registerTemplate(item)
item = Template("Zombie Dust")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\Icons\INV_mined_ironore")
ObjectManagerClient.registerTemplate(item)
item = Template("Healing Potion")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ITEM_potion_A")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK,
AbilityActivateHook("heal potion"))
ObjectManagerClient.registerTemplate(item)
item = Template("Mana Potion")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ITEM_potion_A")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK,
AbilityActivateHook("restore mana potion"))
ObjectManagerClient.registerTemplate(item)
item = Template("Poison Potion")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ITEM_potion_A")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK,
AbilityActivateHook("poison potion"))
ObjectManagerClient.registerTemplate(item)
item = Template("Tome of Heal")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ITEM_book_C")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK,
AbilityActivateHook("teach self heal ability"))
ObjectManagerClient.registerTemplate(item)
item = Template("Tome of Fireball")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ICON,
"Interface\FantasyWorldIcons\ITEM_book_A")
item.put(InventoryClient.ITEM_NAMESPACE, InventoryClient.TEMPL_ACTIVATE_HOOK,
AbilityActivateHook("teach self fireball ability"))
ObjectManagerClient.registerTemplate(item)
#
# Teleporter
#
tmpl = Template("Teleporter")
tmpl.put(WorldManagerClient.NAMESPACE,
WorldManagerClient.TEMPL_DISPLAY_CONTEXT,
DisplayContext("tiny_cube.mesh"))
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_OBJECT_TYPE, ObjectTypes.mob)
tmpl.put(WorldManagerClient.NAMESPACE, WorldManagerClient.TEMPL_NAME, "")
ObjectManagerClient.registerTemplate(tmpl)
| 60.809636 | 121 | 0.66239 | 5,255 | 51,749 | 6.203045 | 0.06432 | 0.035433 | 0.02853 | 0.076449 | 0.88956 | 0.87413 | 0.864497 | 0.827561 | 0.803356 | 0.772034 | 0 | 0.012101 | 0.247889 | 51,749 | 850 | 122 | 60.881176 | 0.825416 | 0.027054 | 0 | 0.643972 | 0 | 0 | 0.158975 | 0.110327 | 0 | 0 | 0 | 0 | 0 | 0 | null | null | 0 | 0.017021 | null | null | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
699509467b4776fdc0b97a22116737e46ae9cf57 | 21,902 | py | Python | Spped Advisory.py | HossseinMoradi/Project4 | dbdbbcd1ef8aded9b86bb23b91657e6f6501806e | [
"MIT"
] | null | null | null | Spped Advisory.py | HossseinMoradi/Project4 | dbdbbcd1ef8aded9b86bb23b91657e6f6501806e | [
"MIT"
] | null | null | null | Spped Advisory.py | HossseinMoradi/Project4 | dbdbbcd1ef8aded9b86bb23b91657e6f6501806e | [
"MIT"
] | null | null | null | import math
import ast
from math import e
from decimal import *
import time
import random
getcontext().prec = 28
# first we have defined functions to access signals and vehicles data
def toList(NestedTuple):
return list(map(toList, NestedTuple)) if isinstance(NestedTuple, (list, tuple)) else NestedTuple
# we want to access vehicles information. In the first part of the below function, we determine the pieces of data that are collected. In the secoond part we determine that which types of vehicles are CV/special (i.e, sending information)
def GetVissimDataVehicles():
global vehsAttributes
global vehsAttNames
vehsAttributesNames = ['No', 'VehType\No', 'Pos', 'VehType\No', 'Lane\Link','DesSpeed', 'Speed', 'DistanceToSigHead','InQueue']
vehsAttributes = toList(Vissim.Net.Vehicles.GetMultipleAttributes(vehsAttributesNames))
vehsAttNames = {}
cnt = 0
for att in vehsAttributesNames:
vehsAttNames.update({att: cnt})
cnt += 1
global vehTypesEquipped
global vehTypesSpecial
vehTypesAttributes = Vissim.Net.VehicleTypes.GetMultipleAttributes(['No', 'IsCV', 'IsSpecial'])
vehTypesEquipped = [x[0] for x in vehTypesAttributes if x[1] == True]
vehTypesSpecial = [x[0] for x in vehTypesAttributes if x[2] == True]
def Signal():
#we define a user attributre to access SimSec
Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('SimSec',Vissim.Net.Simulation.SimulationSecond)
Seconds = Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycSec')
SimSec = Vissim.Net.SignalControllers.ItemByKey(1).AttValue('SimSec')
CLength = 60
GetVissimDataVehicles()
deltaT=1
#we should correlate deltaT with the simulation resolution. In other words, number of simulation per second should be one here.
Starting_time=0
Ending_time=1000
TimeNo=[]
i=Starting_time
k=0
while i< Ending_time:
TimeNo.append(k)
k+=1
i=i+deltaT
# Here we define the following attributes: TimeUntilNextGreen, TimeUntilNextRed, CycleDuration
for i in TimeNo:
if SimSec > (i)*deltaT and SimSec <= (i+1)*deltaT:
# this is anexample of a case where green time durations and cycles are varying at each time step
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('SigState')=='RED':
if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart'):
G1=random.randint(10,17)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenTimeDuration', G1)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenTimeDuration'))
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart') < SimSec :
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')+ Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenTimeDuration'))
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('SigState')=='GREEN':
G1=max(random.randint(10,17),SimSec-Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart'))
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenTimeDuration', G1)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextGreen', -1)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd')-SimSec)
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('SigState')=='RED':
if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart'):
G2=random.randint(10,17)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('GreenTimeDuration', G2)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenTimeDuration'))
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart') < SimSec :
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')+ Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenTimeDuration'))
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('SigState')=='GREEN':
G2=max(random.randint(10,17),SimSec-Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart'))
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('GreenTimeDuration', G2)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextGreen', -1)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd')-SimSec)
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('SigState')=='RED':
if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart'):
G3=random.randint(10,17)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('GreenTimeDuration', G3)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenTimeDuration'))
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart') < SimSec :
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')+ Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenTimeDuration'))
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('SigState')=='GREEN':
G3=max(random.randint(10,17),SimSec-Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart'))
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('GreenTimeDuration', G3)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextGreen', -1)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd')-SimSec)
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('SigState')=='RED':
if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart'):
G4=random.randint(10,17)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('GreenTimeDuration', G4)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenTimeDuration'))
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart') < SimSec :
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextGreen', Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration')+ Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('TimeUntilNextGreen')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenTimeDuration'))
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('SigState')=='GREEN':
G4=max(random.randint(10,17),SimSec-Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart'))
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('GreenTimeDuration', G4)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextGreen', -1)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('TimeUntilNextRed', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd')-SimSec)
Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('CycleDuration',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenTimeDuration')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenTimeDuration')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenTimeDuration')+Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenTimeDuration')+8)
Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('CycleEnd',Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleStart')+Vissim.Net.SignalControllers.ItemByKey(1).AttValue('CycleDuration'))
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd'):
if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd')+2:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenStart', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd')+2)
Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('CycleStart', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart')-2)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenEnd', Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart') + Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenTimeDuration'))
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('GreenStart',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd') + 2)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('GreenEnd',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart') + Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenTimeDuration'))
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('GreenStart',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd') + 2)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('GreenEnd',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart') + Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenTimeDuration'))
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('GreenStart',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd') + 2)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('GreenEnd',Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart') + Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenTimeDuration'))
# you have to make a use of simsec to define variable cycle time
# To this end, we firs, define three attributes in the current signal controller with the names of CycleStart, CycleEnd, CycleDuration.
SimSec = Vissim.Net.SignalControllers.ItemByKey(1).AttValue('SimSec')
#we determine Initial values OF CYCLE START
if SimSec<=1:
Vissim.Net.SignalControllers.ItemByKey(1).SetAttValue('CycleStart', 0)
# When we start the simulation, we determine that the signals are operating upon com script.
if SimSec<=1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('ContrByCOM', True)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('ContrByCOM', True)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('ContrByCOM', True)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('ContrByCOM', True)
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('GreenStart',2)
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenStart') - 1:
if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd') - 1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'GREEN')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED')
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd') - 1:
if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('GreenEnd') + 1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED')
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenStart') - 1:
if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd') - 1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'GREEN')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED')
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd') - 1:
if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).AttValue('GreenEnd') + 1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED')
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenStart') - 1:
if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd') - 1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'GREEN')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED')
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd') - 1:
if SimSec < Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).AttValue('GreenEnd') + 1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED')
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenStart') - 1:
if SimSec <= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd') - 1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'GREEN')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED')
if SimSec >= Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).AttValue('GreenEnd') - 1:
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(2).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(3).SetAttValue('SigState', 'RED')
Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(4).SetAttValue('SigState', 'RED')
def ChangeSpeed():
GetVissimDataVehicles()
MinSpeed=5
MaxSpeed=76
if len(vehsAttributes) > 1:
for vehAttributes in vehsAttributes:
if vehAttributes[vehsAttNames['VehType\\No']] in vehTypesEquipped:
# set easier variables of the current vehicle:
DesSpeed = vehAttributes[vehsAttNames['DesSpeed']]
Speed = vehAttributes[vehsAttNames['Speed']]
DistanceToSigHead = vehAttributes[vehsAttNames['DistanceToSigHead']]
# if the vehicle does not have a upcoming signal: set original desired speed
if DistanceToSigHead <= 0:
vehAttributes[vehsAttNames['DesSpeed']] = MaxSpeed
continue # jump to next vehicle
elif vehAttributes[vehsAttNames['Lane\Link']] == '1':
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('TimeUntilNextGreen')==-1:
vehAttributes[vehsAttNames['DesSpeed']] = MaxSpeed
if Vissim.Net.SignalControllers.ItemByKey(1).SGs.ItemByKey(1).AttValue('TimeUntilNextGreen')!=-1:
vehAttributes[vehsAttNames['DesSpeed']] = MinSpeed
vehicleNumDesiredSpeeds = [[x[vehsAttNames['DesSpeed']], x[vehsAttNames['Speed']]] for x in vehsAttributes]
Vissim.Net.Vehicles.SetMultipleAttributes(('DesSpeed', 'Speed'), vehicleNumDesiredSpeeds)
| 82.338346 | 441 | 0.697516 | 2,333 | 21,902 | 6.546507 | 0.084869 | 0.142736 | 0.297911 | 0.401035 | 0.841616 | 0.841616 | 0.832122 | 0.8187 | 0.804099 | 0.797289 | 0 | 0.024041 | 0.162451 | 21,902 | 265 | 442 | 82.649057 | 0.808548 | 0.051913 | 0 | 0.278075 | 0 | 0 | 0.115152 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | null | 0 | 0.032086 | null | null | 0 | 0 | 0 | 0 | null | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 9 |
69b20446b0c37d914c58cb544ec08ec1b2212bcc | 11,088 | py | Python | skgaip/tspdb/tspdb/src/algorithms/pymf/nmf.py | danielsuo/toy_flood | 471d3c4091d86d4a00fbf910937d4e60fdaf79a1 | [
"MIT"
] | 43 | 2019-12-10T00:05:51.000Z | 2022-03-31T21:21:20.000Z | skgaip/tspdb/tspdb/src/algorithms/pymf/nmf.py | danielsuo/toy_flood | 471d3c4091d86d4a00fbf910937d4e60fdaf79a1 | [
"MIT"
] | 5 | 2021-05-09T01:12:31.000Z | 2022-03-29T17:34:15.000Z | skgaip/tspdb/tspdb/src/algorithms/pymf/nmf.py | danielsuo/toy_flood | 471d3c4091d86d4a00fbf910937d4e60fdaf79a1 | [
"MIT"
] | 14 | 2020-01-13T21:20:07.000Z | 2022-03-31T02:11:26.000Z | # Authors: Christian Thurau
# License: BSD 3 Clause
"""
PyMF Non-negative Matrix Factorization.
NMF: Class for Non-negative Matrix Factorization
[1] Lee, D. D. and Seung, H. S. (1999), Learning the Parts of Objects by Non-negative
Matrix Factorization, Nature 401(6755), 788-799.
"""
import numpy as np
import logging
import logging.config
import scipy.sparse
import scipy.optimize
from cvxopt import solvers, base
from base import PyMFBase
from svd import pinv
__all__ = ["NMF", "RNMF", "NMFALS", "NMFNNLS"]
class NMF(PyMFBase):
"""
NMF(data, num_bases=4)
Non-negative Matrix Factorization. Factorize a data matrix into two matrices
s.t. F = | data - W*H | = | is minimal. H, and W are restricted to non-negative
data. Uses the classicial multiplicative update rule.
Parameters
----------
data : array_like, shape (_data_dimension, _num_samples)
the input data
num_bases: int, optional
Number of bases to compute (column rank of W and row rank of H).
4 (default)
Attributes
----------
W : "data_dimension x num_bases" matrix of basis vectors
H : "num bases x num_samples" matrix of coefficients
ferr : frobenius norm (after calling .factorize())
Example
-------
Applying NMF to some rather stupid data set:
>>> import numpy as np
>>> data = np.array([[1.0, 0.0, 2.0], [0.0, 1.0, 1.0]])
>>> nmf_mdl = NMF(data, num_bases=2, niter=10)
>>> nmf_mdl.factorize()
The basis vectors are now stored in nmf_mdl.W, the coefficients in nmf_mdl.H.
To compute coefficients for an existing set of basis vectors simply copy W
to nmf_mdl.W, and set compute_w to False:
>>> data = np.array([[1.5], [1.2]])
>>> W = np.array([[1.0, 0.0], [0.0, 1.0]])
>>> nmf_mdl = NMF(data, num_bases=2)
>>> nmf_mdl.W = W
>>> nmf_mdl.factorize(niter=20, compute_w=False)
The result is a set of coefficients nmf_mdl.H, s.t. data = W * nmf_mdl.H.
"""
def _update_h(self):
# pre init H1, and H2 (necessary for storing matrices on disk)
H2 = np.dot(np.dot(self.W.T, self.W), self.H) + 10**-9
self.H *= np.dot(self.W.T, self.data[:,:])
self.H /= H2
def _update_w(self):
# pre init W1, and W2 (necessary for storing matrices on disk)
W2 = np.dot(np.dot(self.W, self.H), self.H.T) + 10**-9
self.W *= np.dot(self.data[:,:], self.H.T)
self.W /= W2
self.W /= np.sqrt(np.sum(self.W**2.0, axis=0))
class RNMF(PyMFBase):
"""
RNMF(data, num_bases=4)
Non-negative Matrix Factorization. Factorize a data matrix into two matrices
s.t. F = | data - W*H | = | is minimal. H, and W are restricted to non-negative
values. Uses the classicial multiplicative update rule.
Parameters
----------
data : array_like, shape (_data_dimension, _num_samples)
the input data
num_bases: int, optional
Number of bases to compute (column rank of W and row rank of H).
4 (default)
Attributes
----------
W : "data_dimension x num_bases" matrix of basis vectors
H : "num bases x num_samples" matrix of coefficients
ferr : frobenius norm (after calling .factorize())
Example
-------
Applying NMF to some rather stupid data set:
>>> import numpy as np
>>> data = np.array([[1.0, 0.0, 2.0], [0.0, 1.0, 1.0]])
>>> nmf_mdl = RNMF(data, num_bases=2)
>>> nmf_mdl.factorize()
The basis vectors are now stored in nmf_mdl.W, the coefficients in nmf_mdl.H.
To compute coefficients for an existing set of basis vectors simply copy W
to nmf_mdl.W, and set compute_w to False:
>>> data = np.array([[1.5], [1.2]])
>>> W = np.array([[1.0, 0.0], [0.0, 1.0]])
>>> nmf_mdl = RNMF(data, num_bases=2)
>>> nmf_mdl.W = W
>>> nmf_mdl.factorize(niter=20, compute_w=False)
The result is a set of coefficients nmf_mdl.H, s.t. data = W * nmf_mdl.H.
"""
def __init__(self, data, num_bases=4, lamb=2.0):
# call inherited method
PyMFBase.__init__(self, data, num_bases=num_bases)
self._lamb = lamb
def soft_thresholding(self, X, lamb):
X = np.where(np.abs(X) <= lamb, 0.0, X)
X = np.where(X > lamb, X - lamb, X)
X = np.where(X < -1.0*lamb, X + lamb, X)
return X
def _init_h(self):
self.H = np.random.random((self._num_bases, self._num_samples))
self.H[:,:] = 1.0
# normalized bases
Wnorm = np.sqrt(np.sum(self.W**2.0, axis=0))
self.W /= Wnorm
for i in range(self.H.shape[0]):
self.H[i,:] *= Wnorm[i]
self._update_s()
def _update_s(self):
self.S = self.data - np.dot(self.W, self.H)
self.S = self.soft_thresholding(self.S, self._lamb)
def _update_h(self):
# pre init H1, and H2 (necessary for storing matrices on disk)
H1 = np.dot(self.W.T, self.S - self.data)
H1 = np.abs(H1) - H1
H1 /= (2.0* np.dot(self.W.T, np.dot(self.W, self.H)))
self.H *= H1
# adapt S
self._update_s()
def _update_w(self):
# pre init W1, and W2 (necessary for storing matrices on disk)
W1 = np.dot(self.S - self.data, self.H.T)
#W1 = np.dot(self.data - self.S, self.H.T)
W1 = np.abs(W1) - W1
W1 /= (2.0 * (np.dot(self.W, np.dot(self.H, self.H.T))))
self.W *= W1
class NMFALS(PyMFBase):
"""
NMFALS(data, num_bases=4)
Non-negative Matrix Factorization. Factorize a data matrix into two matrices
s.t. F = | data - W*H | = | is minimal. H, and W are restricted to non-negative
data. Uses the an alternating least squares procedure (quite slow for larger
data sets) and cvxopt, similar to aa.
Parameters
----------
data : array_like, shape (_data_dimension, _num_samples)
the input data
num_bases: int, optional
Number of bases to compute (column rank of W and row rank of H).
4 (default)
Attributes
----------
W : "data_dimension x num_bases" matrix of basis vectors
H : "num bases x num_samples" matrix of coefficients
ferr : frobenius norm (after calling .factorize())
Example
-------
Applying NMF to some rather stupid data set:
>>> import numpy as np
>>> data = np.array([[1.0, 0.0, 2.0], [0.0, 1.0, 1.0]])
>>> nmf_mdl = NMFALS(data, num_bases=2)
>>> nmf_mdl.factorize(niter=10)
The basis vectors are now stored in nmf_mdl.W, the coefficients in nmf_mdl.H.
To compute coefficients for an existing set of basis vectors simply copy W
to nmf_mdl.W, and set compute_w to False:
>>> data = np.array([[1.5], [1.2]])
>>> W = np.array([[1.0, 0.0], [0.0, 1.0]])
>>> nmf_mdl = NMFALS(data, num_bases=2)
>>> nmf_mdl.W = W
>>> nmf_mdl.factorize(niter=1, compute_w=False)
The result is a set of coefficients nmf_mdl.H, s.t. data = W * nmf_mdl.H.
"""
def _update_h(self):
def updatesingleH(i):
# optimize alpha using qp solver from cvxopt
FA = base.matrix(np.float64(np.dot(-self.W.T, self.data[:,i])))
al = solvers.qp(HA, FA, INQa, INQb)
self.H[:,i] = np.array(al['x']).reshape((1,-1))
# float64 required for cvxopt
HA = base.matrix(np.float64(np.dot(self.W.T, self.W)))
INQa = base.matrix(-np.eye(self._num_bases))
INQb = base.matrix(0.0, (self._num_bases,1))
map(updatesingleH, xrange(self._num_samples))
def _update_w(self):
def updatesingleW(i):
# optimize alpha using qp solver from cvxopt
FA = base.matrix(np.float64(np.dot(-self.H, self.data[i,:].T)))
al = solvers.qp(HA, FA, INQa, INQb)
self.W[i,:] = np.array(al['x']).reshape((1,-1))
# float64 required for cvxopt
HA = base.matrix(np.float64(np.dot(self.H, self.H.T)))
INQa = base.matrix(-np.eye(self._num_bases))
INQb = base.matrix(0.0, (self._num_bases,1))
map(updatesingleW, xrange(self._data_dimension))
self.W = self.W/np.sum(self.W, axis=1)
class NMFNNLS(PyMFBase):
"""
NMFNNLS(data, num_bases=4)
Non-negative Matrix Factorization. Factorize a data matrix into two matrices
s.t. F = | data - W*H | = | is minimal. H, and W are restricted to non-negative
data. Uses the Lawsons and Hanson's algorithm for non negative constrained
least squares (-> also see scipy.optimize.nnls)
Parameters
----------
data : array_like, shape (_data_dimension, _num_samples)
the input data
num_bases: int, optional
Number of bases to compute (column rank of W and row rank of H).
4 (default)
Attributes
----------
W : "data_dimension x num_bases" matrix of basis vectors
H : "num bases x num_samples" matrix of coefficients
ferr : frobenius norm (after calling .factorize())
Example
-------
Applying NMF to some rather stupid data set:
>>> import numpy as np
>>> data = np.array([[1.0, 0.0, 2.0], [0.0, 1.0, 1.0]])
>>> nmf_mdl = NMFNNLS(data, num_bases=2)
>>> nmf_mdl.factorize(niter=10)
The basis vectors are now stored in nmf_mdl.W, the coefficients in nmf_mdl.H.
To compute coefficients for an existing set of basis vectors simply copy W
to nmf_mdl.W, and set compute_w to False:
>>> data = np.array([[1.5], [1.2]])
>>> W = np.array([[1.0, 0.0], [0.0, 1.0]])
>>> nmf_mdl = NMFNNLS(data, num_bases=2)
>>> nmf_mdl.W = W
>>> nmf_mdl.factorize(niter=1, compute_w=False)
The result is a set of coefficients nmf_mdl.H, s.t. data = W * nmf_mdl.H.
"""
def _update_h(self):
def updatesingleH(i):
self.H[:,i] = scipy.optimize.nnls(self.W, self.data[:,i])[0]
map(updatesingleH, xrange(self._num_samples))
def _update_w(self):
def updatesingleW(i):
self.W[i,:] = scipy.optimize.nnls(self.H.T, self.data[i,:].T)[0]
map(updatesingleW, xrange(self._data_dimension))
def _test():
import doctest
doctest.testmod()
if __name__ == "__main__":
_test()
| 34.977918 | 86 | 0.554293 | 1,605 | 11,088 | 3.723988 | 0.122118 | 0.040154 | 0.010038 | 0.016731 | 0.80425 | 0.768446 | 0.7477 | 0.734817 | 0.71909 | 0.717417 | 0 | 0.028827 | 0.314845 | 11,088 | 316 | 87 | 35.088608 | 0.757931 | 0.553391 | 0 | 0.285714 | 0 | 0 | 0.007562 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.202381 | false | 0 | 0.107143 | 0 | 0.369048 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
69dbb7ccb70650348502e838d71fddbbb61d1bd0 | 159 | py | Python | starlingx-dashboard/starlingx-dashboard/starlingx_dashboard/horizon/tables/__init__.py | MarioCarrilloA/gui | 03a425f1bc1e4671a110f91eb88c272969dba57c | [
"Apache-2.0"
] | null | null | null | starlingx-dashboard/starlingx-dashboard/starlingx_dashboard/horizon/tables/__init__.py | MarioCarrilloA/gui | 03a425f1bc1e4671a110f91eb88c272969dba57c | [
"Apache-2.0"
] | null | null | null | starlingx-dashboard/starlingx-dashboard/starlingx_dashboard/horizon/tables/__init__.py | MarioCarrilloA/gui | 03a425f1bc1e4671a110f91eb88c272969dba57c | [
"Apache-2.0"
] | null | null | null | from starlingx_dashboard.horizon.tables.actions import FixedWithQueryFilter # noqa
from starlingx_dashboard.horizon.tables.actions import LimitAction # noqa
| 53 | 83 | 0.861635 | 18 | 159 | 7.5 | 0.555556 | 0.192593 | 0.325926 | 0.42963 | 0.711111 | 0.711111 | 0.711111 | 0 | 0 | 0 | 0 | 0 | 0.08805 | 159 | 2 | 84 | 79.5 | 0.931034 | 0.056604 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 0 | 1 | 1 | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 8 |
69ee0718f055cd622b37f02c5907911c4e1c5573 | 2,227 | py | Python | parlai/agents/programr/mappings/storage/brainstorage.py | roholazandie/ParlAI | 32352cab81ecb666aefd596232c5ed9f33cbaeb9 | [
"MIT"
] | null | null | null | parlai/agents/programr/mappings/storage/brainstorage.py | roholazandie/ParlAI | 32352cab81ecb666aefd596232c5ed9f33cbaeb9 | [
"MIT"
] | null | null | null | parlai/agents/programr/mappings/storage/brainstorage.py | roholazandie/ParlAI | 32352cab81ecb666aefd596232c5ed9f33cbaeb9 | [
"MIT"
] | null | null | null | class BrainStorage:
def __init__(self):
pass
def initialize(self):
raise NotImplementedError("Should not call this base method, implementation missing")
def save_properties(self):
raise NotImplementedError("Should not call this base method, implementation missing")
def load_properties(self, collection_name):
raise NotImplementedError("Should not call this base method, implementation missing")
def load_denormal(self, collection_name):
raise NotImplementedError("Should not call this base method, implementation missing")
def load_normal(self, collection_name):
raise NotImplementedError("Should not call this base method, implementation missing")
def load_gender(self, collection_name):
raise NotImplementedError("Should not call this base method, implementation missing")
def load_person(self, collection_name):
raise NotImplementedError("Should not call this base method, implementation missing")
def load_maps(self, collection_name):
raise NotImplementedError("Should not call this base method, implementation missing")
def load_variables(self, collection_name):
raise NotImplementedError("Should not call this base method, implementation missing")
def save_variable(self, collection_name, key, value):
raise NotImplementedError("Should not call this base method, implementation missing")
def save_gender(self, genders):
raise NotImplementedError("Should not call this base method, implementation missing")
def save_normal(self, normals):
raise NotImplementedError("Should not call this base method, implementation missing")
def save_denormal(self, denormals):
raise NotImplementedError("Should not call this base method, implementation missing")
def save_person(self, persons):
raise NotImplementedError("Should not call this base method, implementation missing")
def save_maps(self, maps):
raise NotImplementedError("Should not call this base method, implementation missing")
def save_variables(self, variables):
raise NotImplementedError("Should not call this base method, implementation missing") | 42.826923 | 93 | 0.746295 | 253 | 2,227 | 6.462451 | 0.13834 | 0.234862 | 0.293578 | 0.322936 | 0.850765 | 0.850765 | 0.850765 | 0.850765 | 0.850765 | 0.850765 | 0 | 0 | 0.187247 | 2,227 | 52 | 94 | 42.826923 | 0.903315 | 0 | 0 | 0.457143 | 0 | 0 | 0.402154 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.485714 | false | 0.028571 | 0 | 0 | 0.514286 | 0 | 0 | 0 | 0 | null | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 10 |
384ee6ca346b4215dbad7143cfbf3acefa7f1de5 | 12,033 | py | Python | tests/response_data.py | universalappfactory/pyhearthis | 29c3a7e051d2937ad2fddb48d083e6a48b298014 | [
"MIT"
] | null | null | null | tests/response_data.py | universalappfactory/pyhearthis | 29c3a7e051d2937ad2fddb48d083e6a48b298014 | [
"MIT"
] | null | null | null | tests/response_data.py | universalappfactory/pyhearthis | 29c3a7e051d2937ad2fddb48d083e6a48b298014 | [
"MIT"
] | null | null | null |
WAVEFORM_RESPONSE_DATA = '''[187,191,215,197,211,223,229,236,229,219,188,169,176,167,166,230,240,218,197,216,209,227,229,237,238,245,193,201,208,255,235,242,240,230,212,221,200,207,223,235,235,234,226,188,193,210,229,240,234,222,225,210,202,204,233,237,243,225,220,213,211,215,229,240,241,230,222,209,196,227,220,231,162,161,155,164,188,174,180,169,190,214,216,225,212,177,202,234,227,230,211,216,198,197,196,209,226,231,229,225,206,192,196,210,210,217,230,198,183,173,215,214,199,205,207,207,227,213,207,206,224,200,206,212,223,217,197,187,184,229,212,189,200,214,209,210,215,208,200,219,222,224,187,197,193,180,174,149,158,174,178,187,177,187,191,226,224,203,215,225,216,240,211,219,226,215,212,238,194,215,215,190,193,216,218,235,215,220,225,211,236,226,220,228,243,234,225,219,222,235,217,219,216,224,235,203,213,199,211,206,215,206,213,247,221,218,235,234,216,210,211,238,200,201,217,210,210,195,199,221,198,240,214,223,234,224,235,221,217,201,201,234,210,221,204,197,204,219,212,224,226,213,207,192,246,219,226,210,203,212,203,226,210,208,212,206,213,221,209,209,227,216,225,227,237,212,210,214,228,207,223,216,250,217,199,207,220,228,225,218,204,223,196,208,242,216,212,208,213,233,212,166,168,175,179,169,181,181,177,190,190,203,182,190,192,201,201,207,208,216,217,207,213,212,208,205,204,222,219,216,210,227,239,226,225,219,232,226,214,230,223,244,231,246,234,247,233,244,249,234,226,247,238,242,235,239,210,225,230,246,229,233,247,247,230,243,212,204,192,180,208,182,181,191,167,190,171,186,177,177,185,167,219,240,238,242,230,242,220,227,239,214,245,238,225,241,255,255,225,225,240,242,211,237,234,236,230,237,217,224,234,247,202,217,231,237,228,233,194,200,200,208,204,215,213,212,211,216,209,220,214,207,203,209,217,219,227,227,229,237,228,147,150,147,153,221,232,239,236,235,227,222,225,222,220,222,215,231,228,229,230,231,224,221,234,228,234,229,228,219,200,216,216,214,224,229,235,231,233,238,231,230,228,235,230,222,227,218,194,215,227,239,239,221,226,236,220,227,228,231,228,222,212,201,211,216,204,223,242,242,243,236,218,236,208,229,212,220,206,204,218,224,217,198,212,213,220,250,243,228,227,246,242,249,224,237,241,231,239,216,240,221,214,228,209,237,232,249,232,222,219,215,216,213,244,209,222,239,215,234,220,222,214,210,225,235,244,219,244,233,237,235,248,218,229,236,247,219,243,230,247,241,237,237,236,235,240,241,234,226,242,242,242,237,242,240,247,251,247,248,238,224,242,228,243,212,221,224,226,248,219,211,192,186,200,210,205,192,216,223,218,222,221,249,255,244,233,227,235,249,226,219,220,235,236,230,226,230,234,241,240,244,238,244,225,243,234,244,226,242,228,221,191,181,175,168,163,184,171,183,208,231,232,239,249,226,233,232,224,225,237,246,226,232,238,218,222,226,239,255,248,239,242,244,232,242,229,225,234,225,220,234,218,239,236,235,226,230,214,204,223,202,199,218,208,229,208,227,242,222,237,219,223,217,213,205,222,242,249,224,230,238,238,234,237,221,242,231,206,238,228,235,206,240,217,226,236,208,210,231,244,237,250,231,234,249,251,244,227,238,226,229,212,219,214,228,221,215,208,212,227,217,221,225,219,228,210,237,217,221,197,221,215,219,213,210,223,236,208,222,240,220,233,217,209,226,231,212,232,231,233,210,198,226,225,223,215,217,232,229,229,213,218,234,218,230,240,223,220,224,229,251,221,219,232,229,214,212,234,220,208,211,234,236,217,218,161,163,158,168,227,231,248,225,220,243,236,232,235,243,238,236,222,228,230,229,230,212,235,232,237,219,222,236,223,226,216,234,239,230,224,226,214,210,220,239,229,215,216,224,240,209,226,231,233,214,211,226,230,211,206,223,233,237,200,201,220,216,198,191,203,202,193,236,216,208,207,211,219,210,219,197,224,210,192,226,202,204,205,212,196,206,214,215,188,195,204,202,208,183,199,192,197,195,201,193,224,188,167,176,187,179,192,196,218,207,191,202,221,213,223,211,208,206,212,198,196,208,211,201,192,197,205,203,195,183,207,198,203,186,188,205,205,195,193,216,197,189,189,186,191,188,194,192,198,197,201,186,206,197,203,220,213,213,195,209,212,202,213,210,211,227,184,163,170,157,170,163,170,177,152,167,160,187,168,170,171,180,172,179,187,171,184,215,206,218,216,201,240,215,201,223,202,194,209,206,213,235,208,226,218,210,220,209,214,196,216,223,210,215,190,218,201,216,210,205,210,221,205,216,210,200,209,228,184,215,235,232,227,227,223,227,223,218,231,224,228,222,222,235,229,238,226,227,225,229,236,229,229,241,239,238,240,232,248,232,239,244,235,235,246,242,238,226,233,239,238,230,222,230,225,234,232,231,235,235,242,240,230,226,228,233,230,239,240,244,230,228,245,242,232,233,229,242,237,229,229,245,242,233,233,237,230,231,228,237,242,236,242,236,238,233,240,237,228,232,243,241,238,240,239,232,233,233,237,239,246,226,233,240,238,235,240,239,244,233,237,232,236,248,241,234,232,242,239,239,206,188,192,193,195,193,193,201,206,215,228,228,219,243,229,240,177,212,212,214,214,224,210,218,199,245,214,216,224,219,225,220,229,214,220,234,218,213,218,222,227,234,227,217,224,216,180,163,210,220,208,217,227,226,212,218,220,218,208,223,209,226,210,211,213,217,212,179,171,186,173,188,207,208,204,211,192,200,200,200,207,200,200,211,219,223,241,231,198,221,219,222,195,170,213,226,229,197,199,214,219,224,227,197,197,209,200,212,161,159,202,222,224,223,239,214,200,214,217,212,208,181,166,169,180,216,209,216,221,224,213,218,203,195,230,203,210,195,204,201,207,215,211,203,200,189,200,207,216,207,204,210,201,194,210,219,210,210,190,205,197,189,188,206,184,181,182,191,213,198,204,203,198,203,200,202,190,218,221,193,189,207,207,196,193,208,190,189,226,205,212,200,198,211,198,196,192,208,195,184,187,192,162,159,171,157,168,163,162,153,175,195,216,207,218,210,213,208,228,211,222,202,228,215,209,211,226,224,219,217,224,209,223,212,220,218,221,203,208,224,219,211,204,204,214,221,206,212,200,215,227,214,199,206,204,207,216,209,213,211,215,211,227,218,198,172,158,175,180,181,211,206,181,201,198,224,226,228,220,220,225,224,230,230,215,230,207,214,212,211,215,215,218,233,214,225,207,199,215,220,222,211,209,229,217,214,221,213,189,217,216,215,235,216,218,202,160,169,177,176,162,190,184,171,185,196,195,180,191,229,202,229,221,228,203,213,210,201,211,213,217,215,222,237,237,196,199,209,213,233,229,217,227,222,217,229,228,208,209,219,191,204,211,222,217,192,205,206,198,200,209,212,207,204,210,221,185,200,218,205,209,206,230,200,200,213,196,198,208,198,205,204,172,168,171,177,181,204,205,207,201,201,211,211,206,201,159,213,202,203,202,209,221,166,161,153,159,155,174,162,168,168,179,206,201,201,201,194,210,206,211,190,201,201,192,221,211,212,223,216,216,231,213,174,167,173,167,165,173,172,164,170,163,163,171,177,167,178,172,161,179,190,191,232,212,221,224,220,216,215,241,227,235,208,217,235,208,218,224,215,221,232,220,220,194,202,185,188,215,220,210,212,208,239,226,219,228,221,242,212,230,216,223,227,169,175,216,208,223,233,208,213,228,214,211,209,224,237,210,225,219,226,235,216,227,206,193,175,153,148,167,157,154,215,216,221,229,211,203,215,193,217,195,205,187,224,200,178,196,202,201,226,201,181,203,207,187,175,161,168,168,175,200,235,224,227,220,218,227,207,206,199,209,204,200,211,172,172,162,161,201,202,197,202,194,209,202,233,203,214,209,215,227,198,207,206,180,173,180,178,195,214,212,202,198,201,209,196,205,210,190,206,206,205,229,196,211,209,211,214,202,191,181,185,179,179,189,189,192,199,215,194,207,199,203,202,196,152,151,154,156,160,159,174,165,180,207,199,210,217,210,204,201,204,206,215,200,207,210,223,199,218,229,225,214,237,182,170,171,163,166,157,154,164,162,160,178,169,202,172,167,171,179,175,190,177,185,222,217,199,208,223,217,210,231,220,215,209,221,207,218,221,219,209,214,208,198,212,218,207,212,225,226,209,213,226,230,178,193,197,194,208,188,179,174,177,174,186,182,210,198,217,209,194,178,171,187,178,191,196,197,187,192,205,195,186,174,190,193,178,185,186,195,193,205,189,184,169,179,188,187,207,192,189,204,199,196,177,196,186,178,185,189,188,188,205,210,194,186,177,184,177,183,213,206,207,187,198,192,182,202,202,220,219,201,189,209,198,200,191,186,185,197,195,201,203,195,201,209,199,201,201,187,196,179,188,190,188,205,185,212,191,188,181,184,189,181,190,201,210,219,209,199,182,182,190,217,187,196,196,179,193,211,202,181,191,203,186,194,189,209,198,223,200,204,228,189,215,203,234,210,227,216,220,211,216,217,244,230,223,225,231,243,244,238,231,217,215,237,223,232,220,224,243,217,216,234,233,200,220,228,217,238,226,215,223,227,220,185,182,174,187,223,222,235,235,245,231,201,229,239,222,215,240,223,214,222,214,221,213,221,237,225,222,207,223,230,220,229,222,243,219,216,220,228,223,217,211,232,212,239,217,234,214,222,222,229,214,221,214,215,233,238,236,209,217,217,213,224,224,226,211,221,228,211,213,218,235,210,160,165,162,176,165,211,213,224,237,215,188,182,203,199,200,209,201,214,215,233,205,194,217,251,227,223,214,220,214,214,215,204,209,215,206,200,217,211,203,202,199,198,207,209,213,210,217,240,201,195,197,190,214,210,212,205,208,201,197,203,218,224,241,214,200,205,199,207,198,154,142,156,165,179,163,162,218,207,205,203,200,202,204,200,212,206,201,208,201,202,193,203,200,207,194,194,210,196,204,195,198,208,208,197,202,214,207,203,197,209,203,223,206,193,195,204,191,211,209,214,203,201,213,210,202,215,201,206,195,208,174,173,187,194,185,179,191,199,200,201,207,204,203,193,187,198,202,201,204,176,200,195,193,201,210,206,213,223,218,213,210,201,203,213,211,217,215,221,209,210,187,186,201,197,200,200,164,161,169,165,156,181,180,183,168,172,193,171,193,186,182,198,183,181,176,174,205,200,195,206,187,218,186,192,187,198,217,236,220,227,220,222,218,234,204,173,209,215,213,214,217,217,217,206,220,209,216,221,217,209,195,208,220,215,214,226,227,194,211,179,196,165,176,188,201,199,182,183,190,193,199,203,190,213,207,212,218,214,213,213,205,211,205,160,158,156,173,191,172,178,179,177,184,181,188,187,182,204,241,211,211,212,208,210,197,216,228,203,207,212,206,199,213,202,221,200,206,208,208,213,215,218,234,180,188,191,171,174,178,172,186,181,181,213,216,213,212,210,211,216,204,219,212,217,207,213,196,203,208,219,212,211,218,213,211,212,217,215,214,206,217,219,221,216,224,218,224,232,227,228,232,219,214,227,234,232,234,229,235,227,234,227,237,221,223,230,225,232,247,235,229,230,224,234,232,242,228,227,216,226,226,229,237,228,223,221,229,221,228,225,253,227,232,230,238,236,226,235,230,236,235,229,234,225,229,228,236,246,222,232,222,230,229,230,226,227,231,231,224,220,229,236,234,229,226,226,210,220,215,205,219,218,230,227,219,221,215,224,202,207,207,220,201,216,205,218,211,196,226,230,227,227,235,240,229,208,226,238,252,225,235,244,242,217,227,239,219,210,230,191,191,210,184,183,210,217,229,227,217,205,222,223,208,221,214,229,222,204,221,233,233,217,217,234,211,209,227,222,218,225,219,228,207,188,192,201,181,185,195,192,187,194,218,227,232,223,221,221,228,208,201,224,231,230,208,229,227,224,202,228,230,226,198,226,228,222,208,209,237,229,214,228,231,227,213,204,225,230,186,188,202,195,193,197,202,195,212,217,217,222,225,215,220,209,204,213,218,207,168,183,196,190,240,209,209,210,202,214,207,229,212,206,216,219,211,218,218,219,217,227,221,215,202,209,215,220,215,230,214,221,203,204,210,208,227,208,225,213,196,177,181,198,208,182,190,186,189,198,192,200,202,198,186,192,162,157,165,165,206,211,202,210,214,206,205,226,223,209,213,201,195,204,213,237,208,221,229,191,190,216,212,226,213,217,220,211,208,205,209,199,212,206,216,213,216,212,203,193,208,209,213,215,188,182,203,206,206,219,210,199,202,205,208,203,197,204,203,209,213,190,192,195,220,194,200,215,218,219,199,214,204,196,213,212,213,190,198,208,202,207,205,200,226,212,210,213,207,226,209,202,203,200,214,200,199,237,222,227,218,217,194,211,212,228,233,236,225,227,225,207,209,219,235,234,237,222,219,204,220,236,236,233,229,226,229,219,207,226,218,223,225,223,229,207,200,226,222,219,221,230,232,222,217,231,229,229,230,229,221,231,202,209,216,223,224,230,217,230,213,207,229,224,211,220,229,223,224,194,222,220,231,220,229,228,211,214,211,211,221,228,227,229,208,199,216,230,220,232,218,222,224,213,212,216,223,231,237,220,212,205,207,218,213,220,227,227,232,225,196,212,232,228,216,222,227,217,204,216,213,224,230,235,225,225,215,197,233,207,222,226,235,238,224,185,128]''' | 6,016.5 | 12,032 | 0.749771 | 3,003 | 12,033 | 3.003663 | 0.037629 | 0.004656 | 0.002993 | 0.002661 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.74813 | 0.000249 | 12,033 | 2 | 12,032 | 6,016.5 | 0.001663 | 0 | 0 | 0 | 0 | 1 | 0.997341 | 0.997341 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 |
3861ae18b9f3d8eb13bd13c983b8a0afcb588621 | 183 | py | Python | summarize/data/paragraph_tokenizers/__init__.py | danieldeutsch/summarize | f36a86d58f381ff1f607f356dad3d6ef7b0e0224 | [
"Apache-2.0"
] | 15 | 2019-11-01T11:49:44.000Z | 2021-01-19T06:59:32.000Z | summarize/data/paragraph_tokenizers/__init__.py | CogComp/summary-cloze | b38e3e8c7755903477fd92a4cff27125cbf5553d | [
"Apache-2.0"
] | 2 | 2020-03-30T07:54:01.000Z | 2021-11-15T16:27:42.000Z | summarize/data/paragraph_tokenizers/__init__.py | CogComp/summary-cloze | b38e3e8c7755903477fd92a4cff27125cbf5553d | [
"Apache-2.0"
] | 3 | 2019-12-06T05:57:51.000Z | 2019-12-11T11:34:21.000Z | from summarize.data.paragraph_tokenizers.paragraph_tokenizer import ParagraphTokenizer
from summarize.data.paragraph_tokenizers.paragraph_word_tokenizer import ParagraphWordTokenizer
| 61 | 95 | 0.923497 | 19 | 183 | 8.631579 | 0.526316 | 0.158537 | 0.207317 | 0.317073 | 0.54878 | 0.54878 | 0 | 0 | 0 | 0 | 0 | 0 | 0.043716 | 183 | 2 | 96 | 91.5 | 0.937143 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 7 |
3888f75f351e24884fda5e20895149316b64e35d | 32,931 | py | Python | RouterConfiguration/Juniper/juniper_config_printer.py | nsg-ethz/Metha | 929839994043d80e20151c1bed33f8a1e3c1f8bd | [
"MIT"
] | 9 | 2021-02-24T12:17:28.000Z | 2021-11-29T02:24:58.000Z | RouterConfiguration/Juniper/juniper_config_printer.py | nsg-ethz/Metha | 929839994043d80e20151c1bed33f8a1e3c1f8bd | [
"MIT"
] | null | null | null | RouterConfiguration/Juniper/juniper_config_printer.py | nsg-ethz/Metha | 929839994043d80e20151c1bed33f8a1e3c1f8bd | [
"MIT"
] | 2 | 2021-03-08T22:21:50.000Z | 2021-04-20T06:59:40.000Z | import os
from utils import *
from RouterConfiguration.Juniper.juniper_config_features import *
from network_features import *
feature_config = {
ProtocolIndependentFeatures.STATIC_ROUTE: lambda network,
interface: f'set routing-options static route {int_to_ip(network.address)}/{network.prefix} next-hop {interface.name}',
OSPFFeatures.AREA_RANGE: lambda area, network, override_metric,
restrict: f'set protocols ospf area {area} area-range {int_to_ip(network.address)}/{network.prefix} {override_metric} {restrict}',
OSPFFeatures.AREA_LABEL_SWITCHED_PATH: lambda area, path,
metric: f'set protocols ospf area {area} label-switched-path {path} metric {metric}',
OSPFFeatures.NSSA_DEFAULT_LSA: lambda area, metric, metric_type,
type_7: f'set protocols ospf area {area} nssa default-lsa default-metric {metric} {metric_type} {type_7}',
OSPFFeatures.NSSA_NO_SUMMARIES: lambda area: f'set protocols ospf area {area} nssa no-summaries',
OSPFFeatures.STUB_DEFAULT_METRIC: lambda area,
metric: f'set protocols ospf area {area} stub default-metric {metric}',
OSPFFeatures.STUB_NO_SUMMARIES: lambda area: f'set protocols ospf area {area} stub no-summaries',
OSPFFeatures.EXTERNAL_PREFERENCE: lambda preference: f'set protocols ospf external-preference {preference}',
OSPFFeatures.NO_RFC_1583: lambda: f'set protocols ospf no-rfc-1583',
OSPFFeatures.REFERENCE_BANDWIDTH: lambda bandwidth: f'set protocols ospf reference-bandwidth {bandwidth}',
OSPFFeatures.INTERFACE_LDP_SYNCHRONIZATION: lambda interface,
disable: f'set protocols ospf area {interface.area} interface {interface.name} ldp-sychronization {disable}',
OSPFFeatures.INTERFACE_LINK_PROTECTION: lambda
interface: f'set protocols ospf area {interface.area} interface {interface.name} link-protection',
OSPFFeatures.INTERFACE_METRIC: lambda interface,
metric: f'set protocols ospf area {interface.area} interface {interface.name} metric {metric}',
OSPFFeatures.INTERFACE_PASSIVE: lambda
interface: f'set protocols ospf area {interface.area} interface {interface.name} passive',
OSPFFeatures.INTERFACE_PRIORITY: lambda interface,
priority: f'set protocols ospf area {interface.area} interface {interface.name} priority {priority}',
OSPFFeatures.INTERFACE_TE_METRIC: lambda interface,
metric: f'set protocols ospf area {interface.area} interface {interface.name} te-metric {metric}',
OSPFFeatures.REDISTRIBUTE_DIRECT: lambda: f'set protocols ospf export send-direct',
OSPFFeatures.REDISTRIBUTE_STATIC: lambda: f'set protocols ospf export send-static',
OSPFFeatures.REDISTRIBUTE_BGP: lambda: f'set protocols ospf export send-bgp',
OSPFFeatures.EXPORT: lambda: f'set protocols ospf export export-ospf',
OSPFFeatures.IMPORT: lambda: f'set protocols ospf import import-ospf',
BGPFeatures.ACCEPTED_PREFIX_LIMIT: lambda
limit: f'set protocols bgp family inet any accepted-prefix-limit maximum {limit}',
BGPFeatures.ADVERTISE_EXTERNAL: lambda group: f'set protocols bgp group {group} advertise-external',
BGPFeatures.ADVERTISE_INACTIVE: lambda: f'set protocols bgp advertise-inactive',
BGPFeatures.ADVERTISE_PEER_AS: lambda: f'set protocols bgp advertise-peer-as',
BGPFeatures.AS_OVERRIDE: lambda group: f'set protocols bgp group {group} as-override',
BGPFeatures.CLUSTER: lambda: f'',
BGPFeatures.DAMPING: lambda: f'set protocols bgp damping',
BGPFeatures.ENFORCE_FIRST_AS: lambda: f'set protocols bgp group EBGP enforce-first-as',
BGPFeatures.LOCAL_AS: lambda AS, option: f'set protocols bgp local-as {AS.num} {option}',
BGPFeatures.METRIC_OUT: lambda metric: f'set protocols bgp metric-out {metric}',
BGPFeatures.MULTIHOP: lambda: f'',
BGPFeatures.NO_CLIENT_REFLECT: lambda: f'set protocols bgp no-client-reflect',
BGPFeatures.PASSIVE: lambda: f'set protocols bgp passive',
BGPFeatures.PATH_SELECTION: lambda option: f'set protocols bgp path-selection {option}',
BGPFeatures.REMOVE_PRIVATE: lambda: f'set protocols bgp remove-private',
BGPFeatures.TCP_MSS: lambda size: f'set protocols bgp tcp-mss {size}',
BGPFeatures.ADD_PATH: lambda group,
options: f'set protocols bgp group {group} family inet unicast add-path {options}',
BGPFeatures.LOOPS: lambda loops: f'set protocols bgp family inet unicast loops {loops}',
BGPFeatures.PREFIX_LIMIT: lambda limit: f'set protocols bgp family inet any prefix-limit maximum {limit}',
BGPFeatures.REDISTRIBUTE_DIRECT: lambda: f'set protocols bgp export send-direct',
BGPFeatures.REDISTRIBUTE_STATIC: lambda: f'set protocols bgp export send-static',
BGPFeatures.REDISTRIBUTE_OSPF: lambda: f'set protocols bgp export send-ospf',
BGPFeatures.IMPORT: lambda: f'set protocols bgp import import-bgp',
BGPFeatures.EXPORT: lambda: f'set protocols bgp export export-bgp',
BGPFeatures.LOCAL_PREFERENCE: lambda pref: f'set protocols bgp local-preference {pref}',
BGPFeatures.NEIGHBOUR_POLICY_EXPORT: lambda neighbour, policy: f'set protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} export {policy.name}',
BGPFeatures.NEIGHBOUR_POLICY_IMPORT: lambda neighbour, policy: f'set protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} import {policy.name}',
PolicyFeatures.FROM_AREA: lambda policy, term, area: f'set policy-options policy-statement {policy.name} term {term} from area {area}',
PolicyFeatures.FROM_AS_PATH: lambda policy, term, as_path: f'set policy-options policy-statement {policy.name} term {term} from as-path {as_path}',
PolicyFeatures.FROM_AS_PATH_GROUP: lambda policy, term, group: f'set policy-options policy-statement {policy.name} term {term} from as-path-group {group}',
PolicyFeatures.FROM_COLOR: lambda policy, term, color: f'set policy-options policy-statement {policy.name} term {term} from color {color}',
PolicyFeatures.FROM_COMMUNITY: lambda policy, term, community: f'set policy-options policy-statement {policy.name} term {term} from community {community}',
PolicyFeatures.FROM_FAMILY: lambda policy, term, family: f'set policy-options policy-statement {policy.name} term {term} from family {family}',
PolicyFeatures.FROM_INSTANCE: lambda: f'',
PolicyFeatures.FROM_INTERFACE: lambda policy, term, interface: f'set policy-options policy-statement {policy.name} term {term} from interface {interface.name}',
PolicyFeatures.FROM_LEVEL: lambda: f'',
PolicyFeatures.FROM_LOCAL_PREFERENCE: lambda policy, term, preference: f'set policy-options policy-statement {policy.name} term {term} from local-preference {preference}',
PolicyFeatures.FROM_METRIC: lambda policy, term, metric: f'set policy-options policy-statement {policy.name} term {term} from metric {metric}',
PolicyFeatures.FROM_NEIGHBOUR: lambda policy, term, neighbour: f'set policy-options policy-statement {policy.name} term {term} from neighbor {int_to_ip(neighbour.address)}',
PolicyFeatures.FROM_ORIGIN: lambda policy, term, origin: f'set policy-options policy-statement {policy.name} term {term} from origin {origin}',
PolicyFeatures.FROM_POLICY: lambda policy, term, policy2: f'set policy-options policy-statement {policy.name} term {term} from policy {policy2}',
PolicyFeatures.FROM_PREFIX_LIST: lambda policy, term, prefix_list: f'set policy-options policy-statement {policy.name} term {term} from prefix-list {prefix_list}',
PolicyFeatures.FROM_PREFIX_LIST_FILTER: lambda policy, term, prefix_list, match_type: f'set policy-options policy-statement {policy.name} term {term} from prefix-list-filter {prefix_list} {match_type}',
PolicyFeatures.FROM_PROTOCOL: lambda policy, term, protocol: f'set policy-options policy-statement {policy.name} term {term} from protocol {protocol}',
PolicyFeatures.FROM_RIB: lambda policy, term, rib: f'set policy-options policy-statement {policy.name} term {term} from rib {rib}',
PolicyFeatures.FROM_ROUTE_FILTER: lambda policy, term, net, match_type: f'set policy-options policy-statement {policy.name} term {term} from route-filter {int_to_ip(net.address)}/{net.prefix} {match_type}',
PolicyFeatures.FROM_ROUTE_TYPE: lambda policy, term, route_type: f'set policy-options policy-statement {policy.name} term {term} from route-type {route_type}',
PolicyFeatures.FROM_SOURCE_ADDRESS_FILTER: lambda policy, term, source_filter, match_type: f'set policy-options policy-statement {policy.name} term {term} from source-address-filter {source_filter} {match_type}',
PolicyFeatures.FROM_TAG: lambda policy, term, tag: f'set policy-options policy-statement {policy.name} term {term} from tag {tag}',
PolicyFeatures.FROM_NEXT_HOP: lambda policy, term, next_hop: f'set policy-options policy-statement {policy.name} term {term} from next-hop {int_to_ip(next_hop)}',
PolicyFeatures.TO_LEVEL: lambda: f'',
PolicyFeatures.TO_RIB: lambda: f'',
PolicyFeatures.THEN_ACCEPT: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then accept',
PolicyFeatures.THEN_AS_PATH_EXPAND: lambda policy, term, n: f'set policy-options policy-statement {policy.name} term {term} then as-path-expand last-as count {n}',
PolicyFeatures.THEN_AS_PATH_PREPEND: lambda policy, term, as_path: f'set policy-options policy-statement {policy.name} term {term} then as-path-prepend {as_path}',
PolicyFeatures.THEN_COLOR: lambda policy, term, add, color: f'set policy-options policy-statement {policy.name} term {term} then color {add} {color}',
PolicyFeatures.THEN_COLOR2: lambda policy, term, add, color: f'set policy-options policy-statement {policy.name} term {term} then color2 {add} {color}',
PolicyFeatures.THEN_COMMUNITY_ADD: lambda policy, term, community: f'set policy-options policy-statement {policy.name} term {term} then community add {community}',
PolicyFeatures.THEN_COMMUNITY_DELETE: lambda policy, term, community: f'set policy-options policy-statement {policy.name} term {term} then community delete {community}',
PolicyFeatures.THEN_COMMUNITY_SET: lambda policy, term, community: f'set policy-options policy-statement {policy.name} term {term} then community set {community}',
PolicyFeatures.THEN_COS_NEXT_HOP_MAP: lambda policy, term, cos_map: f'set policy-options policy-statement {policy.name} term {term} then cos-next-hop-map {cos_map}',
PolicyFeatures.THEN_DEFAULT_ACTION_ACCEPT: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then default-action accept',
PolicyFeatures.THEN_DEFAULT_ACTION_REJECT: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then default-action reject',
PolicyFeatures.THEN_EXTERNAL: lambda policy, term, metric_type: f'set policy-options policy-statement {policy.name} term {term} then external type {metric_type}',
PolicyFeatures.THEN_FORWARDING_CLASS: lambda policy, term, fwd_class: f'set policy-options policy-statement {policy.name} term {term} then forwarding-class {fwd_class}',
PolicyFeatures.THEN_INSTALL_NEXTHOP: lambda: f'',
PolicyFeatures.THEN_LOCAL_PREFERENCE: lambda policy, term, pref: f'set policy-options policy-statement {policy.name} term {term} then local-preference {pref}',
PolicyFeatures.THEN_METRIC: lambda policy, term, metric: f'set policy-options policy-statement {policy.name} term {term} then metric {metric}',
PolicyFeatures.THEN_METRIC_ADD: lambda policy, term, metric: f'set policy-options policy-statement {policy.name} term {term} then metric add {metric}',
PolicyFeatures.THEN_METRIC_EXPRESSION: lambda: f'',
PolicyFeatures.THEN_METRIC_IGP: lambda policy, term, offset: f'set policy-options policy-statement {policy.name} term {term} then metric igp {offset}',
PolicyFeatures.THEN_METRIC2: lambda policy, term, metric: f'set policy-options policy-statement {policy.name} term {term} then metric2 {metric}',
PolicyFeatures.THEN_METRIC2_EXPRESSION: lambda: f'',
PolicyFeatures.THEN_NEXT_HOP: lambda policy, term, next_hop: f'set policy-options policy-statement {policy.name} term {term} then next-hop {int_to_ip(next_hop)}',
PolicyFeatures.THEN_NEXT_HOP_SELF: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then next-hop self',
PolicyFeatures.THEN_NEXT_POLICY: lambda: f'',
PolicyFeatures.THEN_NEXT_TERM: lambda: f'',
PolicyFeatures.THEN_ORIGIN: lambda policy, term, origin: f'set policy-options policy-statement {policy.name} term {term} then origin {origin}',
PolicyFeatures.THEN_PREFERENCE: lambda policy, term, pref: f'set policy-options policy-statement {policy.name} term {term} then preference {pref}',
PolicyFeatures.THEN_PRIORITY: lambda policy, term, priority: f'set policy-options policy-statement {policy.name} term {term} then priority {priority}',
PolicyFeatures.THEN_REJECT: lambda policy, term: f'set policy-options policy-statement {policy.name} term {term} then reject',
PolicyFeatures.THEN_TAG: lambda policy, term, tag: f'set policy-options policy-statement {policy.name} term {term} then tag {tag}',
PolicyFeatures.POLICY_MATCH_FEATURE_BGP_OUT: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args),
PolicyFeatures.POLICY_MATCH_FEATURE_BGP_IN: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args),
PolicyFeatures.POLICY_MATCH_FEATURE_OSPF_OUT: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args),
PolicyFeatures.POLICY_MATCH_FEATURE_OSPF_IN: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args),
PolicyFeatures.POLICY_SET_FEATURE_BGP_OUT: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args),
PolicyFeatures.POLICY_SET_FEATURE_BGP_IN: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args),
PolicyFeatures.POLICY_SET_FEATURE_OSPF_OUT: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args),
PolicyFeatures.POLICY_SET_FEATURE_OSPF_IN: lambda policy, term, feature, *args: feature_config[feature](policy, term, *args)
}
feature_disable = {
ProtocolIndependentFeatures.STATIC_ROUTE: lambda network,
interface: f'delete routing-options static route {int_to_ip(network.address)}/{network.prefix} next-hop {interface.name}',
OSPFFeatures.AREA_RANGE: lambda area, network, override_metric,
restrict: f'delete protocols ospf area {area} area-range {int_to_ip(network.address)}/{network.prefix}',
OSPFFeatures.AREA_LABEL_SWITCHED_PATH: lambda area, path,
metric: f'delete protocols ospf area {area} label-switched-path {path} metric {metric}',
OSPFFeatures.NSSA_DEFAULT_LSA: lambda area, metric, metric_type,
type_7: f'delete protocols ospf area {area} nssa default-lsa',
OSPFFeatures.NSSA_NO_SUMMARIES: lambda area: f'delete protocols ospf area {area} nssa no-summaries',
OSPFFeatures.STUB_DEFAULT_METRIC: lambda area,
metric: f'delete protocols ospf area {area} stub default-metric {metric}',
OSPFFeatures.STUB_NO_SUMMARIES: lambda area: f'delete protocols ospf area {area} stub no-summaries',
OSPFFeatures.EXTERNAL_PREFERENCE: lambda preference: f'delete protocols ospf external-preference {preference}',
OSPFFeatures.NO_RFC_1583: lambda: f'delete protocols ospf no-rfc-1583',
OSPFFeatures.REFERENCE_BANDWIDTH: lambda bandwidth: f'delete protocols ospf reference-bandwidth {bandwidth}',
OSPFFeatures.INTERFACE_LDP_SYNCHRONIZATION: lambda interface,
disable: f'delete protocols ospf area {interface.area} interface {interface.name} ldp-sychronization',
OSPFFeatures.INTERFACE_LINK_PROTECTION: lambda
interface: f'delete protocols ospf area {interface.area} interface {interface.name} link-protection',
OSPFFeatures.INTERFACE_METRIC: lambda interface,
metric: f'delete protocols ospf area {interface.area} interface {interface.name} metric {metric}',
OSPFFeatures.INTERFACE_PASSIVE: lambda
interface: f'delete protocols ospf area {interface.area} interface {interface.name} passive',
OSPFFeatures.INTERFACE_PRIORITY: lambda interface,
priority: f'delete protocols ospf area {interface.area} interface {interface.name} priority {priority}',
OSPFFeatures.INTERFACE_TE_METRIC: lambda interface,
metric: f'delete protocols ospf area {interface.area} interface {interface.name} te-metric {metric}',
OSPFFeatures.REDISTRIBUTE_DIRECT: lambda: f'delete protocols ospf export send-direct',
OSPFFeatures.REDISTRIBUTE_STATIC: lambda: f'delete protocols ospf export send-static',
OSPFFeatures.REDISTRIBUTE_BGP: lambda: f'delete protocols ospf export send-bgp',
OSPFFeatures.EXPORT: lambda: f'delete protocols ospf export export-ospf',
OSPFFeatures.IMPORT: lambda: f'delete protocols ospf import import-ospf',
BGPFeatures.ACCEPTED_PREFIX_LIMIT: lambda
limit: f'delete protocols bgp family inet any accepted-prefix-limit maximum {limit}',
BGPFeatures.ADVERTISE_EXTERNAL: lambda group: f'delete protocols bgp group {group} advertise-external',
BGPFeatures.ADVERTISE_INACTIVE: lambda: f'delete protocols bgp advertise-inactive',
BGPFeatures.ADVERTISE_PEER_AS: lambda: f'delete protocols bgp advertise-peer-as',
BGPFeatures.AS_OVERRIDE: lambda group: f'delete protocols bgp group {group} as-override',
BGPFeatures.CLUSTER: lambda: f'',
BGPFeatures.DAMPING: lambda: f'delete protocols bgp damping',
BGPFeatures.ENFORCE_FIRST_AS: lambda: f'delete protocols bgp group EBGP enforce-first-as',
BGPFeatures.LOCAL_AS: lambda AS, option: f'delete protocols bgp local-as {AS.num}',
BGPFeatures.METRIC_OUT: lambda metric: f'delete protocols bgp metric-out {metric}',
BGPFeatures.MULTIHOP: lambda: f'',
BGPFeatures.NO_CLIENT_REFLECT: lambda: f'delete protocols bgp no-client-reflect',
BGPFeatures.PASSIVE: lambda: f'delete protocols bgp passive',
BGPFeatures.PATH_SELECTION: lambda option: f'delete protocols bgp path-selection',
BGPFeatures.REMOVE_PRIVATE: lambda: f'delete protocols bgp remove-private',
BGPFeatures.TCP_MSS: lambda size: f'delete protocols bgp tcp-mss {size}',
BGPFeatures.ADD_PATH: lambda group, options: f'delete protocols bgp group {group} family inet unicast add-path',
BGPFeatures.LOOPS: lambda loops: f'delete protocols bgp family inet unicast loops {loops}',
BGPFeatures.PREFIX_LIMIT: lambda limit: f'delete protocols bgp family inet any prefix-limit maximum {limit}',
BGPFeatures.REDISTRIBUTE_DIRECT: lambda: f'delete protocols bgp export send-direct',
BGPFeatures.REDISTRIBUTE_STATIC: lambda: f'delete protocols bgp export send-static',
BGPFeatures.REDISTRIBUTE_OSPF: lambda: f'delete protocols bgp export send-ospf',
BGPFeatures.IMPORT: lambda: f'delete protocols bgp import import-bgp',
BGPFeatures.EXPORT: lambda: f'delete protocols bgp export export-bgp',
BGPFeatures.LOCAL_PREFERENCE: lambda pref: f'delete protocols bgp local-preference {pref}',
BGPFeatures.NEIGHBOUR_POLICY_EXPORT: lambda neighbour, policy: f'delete protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} export {policy.name}',
BGPFeatures.NEIGHBOUR_POLICY_IMPORT: lambda neighbour, policy: f'delete protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} import {policy.name}',
PolicyFeatures.FROM_AREA: lambda policy, term, area: f'delete policy-options policy-statement {policy.name} term {term} from area {area}',
PolicyFeatures.FROM_AS_PATH: lambda policy, term, as_path: f'delete policy-options policy-statement {policy.name} term {term} from as-path {as_path}',
PolicyFeatures.FROM_AS_PATH_GROUP: lambda policy, term, group: f'delete policy-options policy-statement {policy.name} term {term} from as-path-group {group}',
PolicyFeatures.FROM_COLOR: lambda policy, term, color: f'delete policy-options policy-statement {policy.name} term {term} from color {color}',
PolicyFeatures.FROM_COMMUNITY: lambda policy, term, community: f'delete policy-options policy-statement {policy.name} term {term} from community {community}',
PolicyFeatures.FROM_FAMILY: lambda policy, term, family: f'delete policy-options policy-statement {policy.name} term {term} from family {family}',
PolicyFeatures.FROM_INSTANCE: lambda: f'',
PolicyFeatures.FROM_INTERFACE: lambda policy, term, interface: f'delete policy-options policy-statement {policy.name} term {term} from interface {interface.name}',
PolicyFeatures.FROM_LEVEL: lambda: f'',
PolicyFeatures.FROM_LOCAL_PREFERENCE: lambda policy, term, preference: f'delete policy-options policy-statement {policy.name} term {term} from local-preference {preference}',
PolicyFeatures.FROM_METRIC: lambda policy, term, metric: f'delete policy-options policy-statement {policy.name} term {term} from metric {metric}',
PolicyFeatures.FROM_NEIGHBOUR: lambda policy, term, neighbour: f'delete policy-options policy-statement {policy.name} term {term} from neighbor {int_to_ip(neighbour.address)}',
PolicyFeatures.FROM_ORIGIN: lambda policy, term, origin: f'delete policy-options policy-statement {policy.name} term {term} from origin {origin}',
PolicyFeatures.FROM_POLICY: lambda policy, term, policy2: f'delete policy-options policy-statement {policy.name} term {term} from policy {policy2}',
PolicyFeatures.FROM_PREFIX_LIST: lambda policy, term, prefix_list: f'delete policy-options policy-statement {policy.name} term {term} from prefix-list {prefix_list}',
PolicyFeatures.FROM_PREFIX_LIST_FILTER: lambda policy, term, prefix_list, match_type: f'delete policy-options policy-statement {policy.name} term {term} from prefix-list-filter {prefix_list} {match_type}',
PolicyFeatures.FROM_PROTOCOL: lambda policy, term, protocol: f'delete policy-options policy-statement {policy.name} term {term} from protocol {protocol}',
PolicyFeatures.FROM_RIB: lambda policy, term, rib: f'delete policy-options policy-statement {policy.name} term {term} from rib {rib}',
PolicyFeatures.FROM_ROUTE_FILTER: lambda policy, term, net, match_type: f'delete policy-options policy-statement {policy.name} term {term} from route-filter {int_to_ip(net.address)}/{net.prefix} {match_type}',
PolicyFeatures.FROM_ROUTE_TYPE: lambda policy, term, route_type: f'delete policy-options policy-statement {policy.name} term {term} from route-type {route_type}',
PolicyFeatures.FROM_SOURCE_ADDRESS_FILTER: lambda policy, term, source_filter, match_type: f'delete policy-options policy-statement {policy.name} term {term} from source-address-filter {source_filter} {match_type}',
PolicyFeatures.FROM_TAG: lambda policy, term, tag: f'delete policy-options policy-statement {policy.name} term {term} from tag {tag}',
PolicyFeatures.FROM_NEXT_HOP: lambda policy, term, next_hop: f'delete policy-options policy-statement {policy.name} term {term} from next-hop {int_to_ip(next_hop)}',
PolicyFeatures.TO_LEVEL: lambda: f'',
PolicyFeatures.TO_RIB: lambda: f'',
PolicyFeatures.THEN_ACCEPT: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then accept',
PolicyFeatures.THEN_AS_PATH_EXPAND: lambda policy, term, n: f'delete policy-options policy-statement {policy.name} term {term} then as-path-expand last-as count {n}',
PolicyFeatures.THEN_AS_PATH_PREPEND: lambda policy, term, as_path: f'delete policy-options policy-statement {policy.name} term {term} then as-path-prepend {as_path}',
PolicyFeatures.THEN_COLOR: lambda policy, term, add, color: f'delete policy-options policy-statement {policy.name} term {term} then color {add} {color}',
PolicyFeatures.THEN_COLOR2: lambda policy, term, add, color: f'delete policy-options policy-statement {policy.name} term {term} then color2 {add} {color}',
PolicyFeatures.THEN_COMMUNITY_ADD: lambda policy, term, community: f'delete policy-options policy-statement {policy.name} term {term} then community add {community}',
PolicyFeatures.THEN_COMMUNITY_DELETE: lambda policy, term, community: f'delete policy-options policy-statement {policy.name} term {term} then community delete {community}',
PolicyFeatures.THEN_COMMUNITY_SET: lambda policy, term, community: f'delete policy-options policy-statement {policy.name} term {term} then community set {community}',
PolicyFeatures.THEN_COS_NEXT_HOP_MAP: lambda policy, term, cos_map: f'delete policy-options policy-statement {policy.name} term {term} then cos-next-hop-map {cos_map}',
PolicyFeatures.THEN_DEFAULT_ACTION_ACCEPT: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then default-action accept',
PolicyFeatures.THEN_DEFAULT_ACTION_REJECT: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then default-action reject',
PolicyFeatures.THEN_EXTERNAL: lambda policy, term, metric_type: f'delete policy-options policy-statement {policy.name} term {term} then external type {metric_type}',
PolicyFeatures.THEN_FORWARDING_CLASS: lambda policy, term, fwd_class: f'delete policy-options policy-statement {policy.name} term {term} then forwarding-class {fwd_class}',
PolicyFeatures.THEN_INSTALL_NEXTHOP: lambda: f'',
PolicyFeatures.THEN_LOCAL_PREFERENCE: lambda policy, term, pref: f'delete policy-options policy-statement {policy.name} term {term} then local-preference {pref}',
PolicyFeatures.THEN_METRIC: lambda policy, term, metric: f'delete policy-options policy-statement {policy.name} term {term} then metric {metric}',
PolicyFeatures.THEN_METRIC_ADD: lambda policy, term, metric: f'delete policy-options policy-statement {policy.name} term {term} then metric add {metric}',
PolicyFeatures.THEN_METRIC_EXPRESSION: lambda: f'',
PolicyFeatures.THEN_METRIC_IGP: lambda policy, term, offset: f'delete policy-options policy-statement {policy.name} term {term} then metric igp {offset}',
PolicyFeatures.THEN_METRIC2: lambda policy, term, metric: f'delete policy-options policy-statement {policy.name} term {term} then metric2 {metric}',
PolicyFeatures.THEN_METRIC2_EXPRESSION: lambda: f'',
PolicyFeatures.THEN_NEXT_HOP: lambda policy, term, next_hop: f'delete policy-options policy-statement {policy.name} term {term} then next-hop {int_to_ip(next_hop)}',
PolicyFeatures.THEN_NEXT_HOP_SELF: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then next-hop self',
PolicyFeatures.THEN_NEXT_POLICY: lambda: f'',
PolicyFeatures.THEN_NEXT_TERM: lambda: f'',
PolicyFeatures.THEN_ORIGIN: lambda policy, term, origin: f'delete policy-options policy-statement {policy.name} term {term} then origin {origin}',
PolicyFeatures.THEN_PREFERENCE: lambda policy, term, pref: f'delete policy-options policy-statement {policy.name} term {term} then preference {pref}',
PolicyFeatures.THEN_PRIORITY: lambda policy, term, priority: f'delete policy-options policy-statement {policy.name} term {term} then priority {priority}',
PolicyFeatures.THEN_REJECT: lambda policy, term: f'delete policy-options policy-statement {policy.name} term {term} then reject',
PolicyFeatures.THEN_TAG: lambda policy, term, tag: f'delete policy-options policy-statement {policy.name} term {term} then tag {tag}',
PolicyFeatures.POLICY_MATCH_FEATURE_BGP_OUT: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args),
PolicyFeatures.POLICY_MATCH_FEATURE_BGP_IN: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args),
PolicyFeatures.POLICY_MATCH_FEATURE_OSPF_OUT: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args),
PolicyFeatures.POLICY_MATCH_FEATURE_OSPF_IN: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args),
PolicyFeatures.POLICY_SET_FEATURE_BGP_OUT: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args),
PolicyFeatures.POLICY_SET_FEATURE_BGP_IN: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args),
PolicyFeatures.POLICY_SET_FEATURE_OSPF_OUT: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args),
PolicyFeatures.POLICY_SET_FEATURE_OSPF_IN: lambda policy, term, feature, *args: feature_disable[feature](policy, term, *args)
}
def config_mode(router, feature, arg):
return ()
def exit_config_mode(feature):
return []
def generate_maps_lists_config(router):
config = []
for as_path in router.as_path_lists:
config.append(f'set policy-options as-path {as_path.name} {as_path.regex}')
for community in router.comm_lists:
config.append(f'set policy-options community {community.name} members [{",".join(community.comms)}]')
for prefix_list in router.prefix_lists:
for seq in prefix_list.prefix:
net = prefix_list.prefix[seq]
config.append(f'set policy-options prefix-list {prefix_list.name} {int_to_ip(net.address)}/{net.prefix}')
for policy in router.ospf_in_route_maps:
config.append(f'set policy-options policy-statement {policy.name} then accept')
for policy in router.ospf_out_route_maps:
config.append(f'set policy-options policy-statement {policy.name} then accept')
for policy in router.bgp_in_route_maps:
config.append(f'set policy-options policy-statement {policy.name} then accept')
for policy in router.bgp_out_route_maps:
config.append(f'set policy-options policy-statement {policy.name} then accept')
return config
def generate_bgp_config(router):
config = [
f'set routing-options autonomous-system {router.AS.num}',
f'set routing-options router-id {int_to_ip(router.router_id)}',
f'set protocols bgp group IBGP type internal',
f'set protocols bgp group IBGP local-address {int_to_ip(router.router_id)}',
f'set protocols bgp group EBGP type external'
]
for neighbour in router.bgp_neighbours:
if neighbour.AS == router.AS:
config.append(f'set protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)}')
else:
config.append(
f'set protocols bgp group {neighbour.group} neighbor {int_to_ip(neighbour.address)} peer-as {neighbour.AS.num}')
return config
def get_base_config(router):
config = [
f'set system host-name {router.name}',
f'set interfaces lo0.0 family inet address {int_to_ip(router.router_id)}/32'
]
for interface in router.interfaces:
if interface.address is not None:
config.append(f'set interfaces {interface.name} family inet address '
f'{int_to_ip(interface.address)}/{interface.prefix}')
if interface.area is not None:
config.append(f'set protocols ospf area {interface.area} interface {interface.name}')
for network, interface in router.fixed_static_routes:
config.append(f'set routing-options static route {int_to_ip(network.address)}/{network.prefix} next-hop {interface.name}')
for area in router.ospf_areas:
if area.type == OSPF_Area_Type.STUB:
config.append(f'set protocols ospf area {area} stub')
elif area.type == OSPF_Area_Type.NSSA:
config.append(f'set protocols ospf area {area} nssa')
for protocol in ['direct', 'static', 'ospf', 'bgp']:
config.append(f'set policy-options policy-statement send-{protocol} term 1 from protocol {protocol}')
config.append(f'set policy-options policy-statement send-{protocol} term 1 then accept')
if Protocols.BGP in router.enabled_protocols:
config.append(f'set protocols bgp hold-time 3')
config.extend(generate_bgp_config(router))
config.extend(generate_maps_lists_config(router))
return config
def write_config(router, path):
config = get_base_config(router)
os.makedirs(os.path.dirname(path), exist_ok=True)
with open(f'{path}{router.name}.set', 'w') as f:
f.write('\n'.join(config))
| 89.486413 | 219 | 0.745832 | 4,332 | 32,931 | 5.543167 | 0.045937 | 0.051639 | 0.071961 | 0.114271 | 0.94453 | 0.925457 | 0.91388 | 0.890684 | 0.886853 | 0.852622 | 0 | 0.001431 | 0.151286 | 32,931 | 367 | 220 | 89.730245 | 0.857741 | 0 | 0 | 0.169753 | 0 | 0.317901 | 0.455498 | 0.028393 | 0 | 0 | 0 | 0 | 0 | 1 | 0.018519 | false | 0.018519 | 0.030864 | 0.006173 | 0.064815 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
2a34900f1755015a41fdc6669c2a89e96570d4f2 | 9,790 | py | Python | tests/test_selective_extraction.py | alizab223/swish-utilities | cc3c30c3aad4fd40f7a72f939b92234d3acf4be2 | [
"Apache-2.0"
] | null | null | null | tests/test_selective_extraction.py | alizab223/swish-utilities | cc3c30c3aad4fd40f7a72f939b92234d3acf4be2 | [
"Apache-2.0"
] | null | null | null | tests/test_selective_extraction.py | alizab223/swish-utilities | cc3c30c3aad4fd40f7a72f939b92234d3acf4be2 | [
"Apache-2.0"
] | null | null | null | import json
from unittest import TestCase
from click.testing import CliRunner
from pandas.io.parsers import read_csv
import requests_mock
from run import cli, main
from tests.common import SNOW_RESPONSE1, SNOW_RESPONSE_WITH_CUSTOM_ID, UNITEST_OUTPUT_FILE, UNITEST_OUTPUT_FILE_CSV, UNITEST_OUTPUT_FILE_PREFIX, patch_for_tests
import os
import os.path
import csv
patch_for_tests()
TEST_IDS_FILE = "tests/data/sys_ids_list.csv"
TEST_IDS_FILE2 = "tests/data/custom_ids_list.csv"
class SelectiveExtractionTesting(TestCase):
def setUp(self) -> None:
if os.path.isfile(UNITEST_OUTPUT_FILE):
os.remove(UNITEST_OUTPUT_FILE)
if os.path.isfile(UNITEST_OUTPUT_FILE_CSV):
os.remove(UNITEST_OUTPUT_FILE_CSV)
def test_no_filter(self):
assert not os.path.isfile(
UNITEST_OUTPUT_FILE), "The file should be deleted"
mock_session = requests_mock.Mocker()
mock_session.register_uri(requests_mock.ANY,
'https://dev71074.service-now.com/api/now/table/sys_audit',
text=SNOW_RESPONSE1)
mock_session.start()
args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident",
"--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000",
"--start_date", "2021-10-03", "--end_date", "2021-10-04"]
runner = CliRunner()
result = runner.invoke(cli, args, catch_exceptions=False)
print(result.output)
assert result.exit_code == 0
assert os.path.isfile(UNITEST_OUTPUT_FILE), "No output file found"
data = []
with open(UNITEST_OUTPUT_FILE, 'r') as f:
data = json.load(f)
assert len([d for d in data if d['sys_id'] ==
'669009b4874330105fd965f73cbb3533'])
assert len([d for d in data if d['sys_id'] == '123'])
assert len([d for d in data if d['sys_id'] == '456'])
assert not len([d for d in data if d['sys_id'] == '567'])
def test_no_filter_to_csv(self):
assert not os.path.isfile(
UNITEST_OUTPUT_FILE_CSV), "The file should be deleted"
mock_session = requests_mock.Mocker()
mock_session.register_uri(requests_mock.ANY,
'https://dev71074.service-now.com/api/now/table/sys_audit',
text=SNOW_RESPONSE1)
mock_session.start()
args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident",
"--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000",
"--start_date", "2021-10-03", "--end_date", "2021-10-04", "--output_format", "csv"]
runner = CliRunner()
result = runner.invoke(cli, args, catch_exceptions=False)
print(result.output)
assert result.exit_code == 0
assert os.path.isfile(UNITEST_OUTPUT_FILE_CSV), "No output file found"
data = []
with open(UNITEST_OUTPUT_FILE_CSV, 'r') as f:
data = list(csv.DictReader(f))
assert len([d for d in data if d['sys_id'] ==
'669009b4874330105fd965f73cbb3533'])
assert len([d for d in data if d['sys_id'] == '123'])
assert len([d for d in data if d['sys_id'] == '456'])
assert not len([d for d in data if d['sys_id'] == '567'])
def test_selective(self):
assert not os.path.isfile(
UNITEST_OUTPUT_FILE), "The file should be deleted"
mock_session = requests_mock.Mocker()
mock_session.register_uri(requests_mock.ANY,
'https://dev71074.service-now.com/api/now/table/sys_audit',
text=SNOW_RESPONSE1)
mock_session.start()
args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident",
"--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000",
"--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", TEST_IDS_FILE]
runner = CliRunner()
result = runner.invoke(cli, args, catch_exceptions=False)
print(result.output)
assert result.exit_code == 0
assert os.path.isfile(
UNITEST_OUTPUT_FILE), "No output file found"
data = []
with open(UNITEST_OUTPUT_FILE, 'r') as f:
data = json.load(f)
assert len([d for d in data if d['sys_id'] ==
'669009b4874330105fd965f73cbb3533']) == 0
ids = read_csv(TEST_IDS_FILE, encoding='utf-8')['sys_id']
for id in ids:
assert len([d for d in data if d['sys_id'] == str(id)])
def test_selective_with_prop_extraction(self):
assert not os.path.isfile(
UNITEST_OUTPUT_FILE), "The file should be deleted"
mock_session = requests_mock.Mocker()
mock_session.register_uri(requests_mock.ANY,
'https://dev71074.service-now.com/api/now/table/sys_audit',
text=SNOW_RESPONSE1)
mock_session.start()
args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident",
"--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000",
"--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", TEST_IDS_FILE,
"--out_props_csv_path", "qwerty.csv", "--out_prop_name", "sys_created_on"]
runner = CliRunner()
result = runner.invoke(cli, args, catch_exceptions=False)
print(result.output)
assert result.exit_code == 0
data = read_csv(f"qwerty.csv", encoding='utf-8')
assert len(data['sys_created_on'].values) == 1, "wrong output count"
def test_selective_with_id_column(self):
assert not os.path.isfile(
UNITEST_OUTPUT_FILE), "The file should be deleted"
mock_session = requests_mock.Mocker()
mock_session.register_uri(requests_mock.ANY,
'https://dev71074.service-now.com/api/now/table/sys_audit',
text=SNOW_RESPONSE_WITH_CUSTOM_ID)
mock_session.start()
args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident",
"--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000",
"--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", TEST_IDS_FILE2,
"--id_field_name", "custom_sys_id"]
runner = CliRunner()
result = runner.invoke(cli, args, catch_exceptions=False)
print(result.output)
assert result.exit_code == 0
assert os.path.isfile(UNITEST_OUTPUT_FILE), "No output file found"
data = []
with open(UNITEST_OUTPUT_FILE, 'r') as f:
data = json.load(f)
assert len([d for d in data if d['custom_sys_id'] ==
'669009b4874330105fd965f73cbb3533']) == 0
ids = read_csv(TEST_IDS_FILE2, encoding='utf-8')['custom_sys_id']
for id in ids:
assert len([d for d in data if d['custom_sys_id'] == str(id)])
def test_selective_with_data_id_name(self):
assert not os.path.isfile(
UNITEST_OUTPUT_FILE), "The file should be deleted"
mock_session = requests_mock.Mocker()
mock_session.register_uri(requests_mock.ANY,
'https://dev71074.service-now.com/api/now/table/sys_audit',
text=SNOW_RESPONSE_WITH_CUSTOM_ID)
mock_session.start()
args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident",
"--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000",
"--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", TEST_IDS_FILE2,
"--id_field_name", "custom_sys_id", "--data_id_name", "custom_sys_id"]
runner = CliRunner()
result = runner.invoke(cli, args, catch_exceptions=False)
print(result.output)
assert result.exit_code == 0
assert os.path.isfile(UNITEST_OUTPUT_FILE), "No output file found"
data = []
with open(UNITEST_OUTPUT_FILE, 'r') as f:
data = json.load(f)
assert len([d for d in data if d['custom_sys_id'] ==
'669009b4874330105fd965f73cbb3533']) == 0
ids = read_csv(TEST_IDS_FILE, encoding='utf-8')['sys_id']
for id in ids:
assert len([d for d in data if d['custom_sys_id'] == str(id)])
def test_file_doesnt_exist(self):
args = ["--extract", "--url", "https://dev71074.service-now.com/api/now/table/sys_audit?sysparm_query=tablename=incident",
"--username", "fake_user", "--password", "fake_pass", "--batch_size", "10000", "--file_limit", "50000",
"--start_date", "2021-10-03", "--end_date", "2021-10-04", "--id_list_path", "qwqweqweqweqew.txt"]
runner = CliRunner()
result = runner.invoke(cli, args, catch_exceptions=False)
print(result.output)
assert result.exit_code == 0
assert not os.path.isfile(
UNITEST_OUTPUT_FILE), "The file should not exist"
| 47.067308 | 160 | 0.60143 | 1,252 | 9,790 | 4.46885 | 0.115016 | 0.051832 | 0.072922 | 0.047542 | 0.866488 | 0.853262 | 0.853262 | 0.840572 | 0.83378 | 0.826273 | 0 | 0.057064 | 0.258938 | 9,790 | 207 | 161 | 47.294686 | 0.714128 | 0 | 0 | 0.735294 | 0 | 0.041176 | 0.288355 | 0.022165 | 0 | 0 | 0 | 0 | 0.2 | 1 | 0.047059 | false | 0.041176 | 0.058824 | 0 | 0.111765 | 0.041176 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
2a994de539e2b5e6ddbcd4033437ddb17e4fe853 | 5,297 | py | Python | main.py | Kubik-Kubik/cs-go-map-chooser | 82e2e27999d151a06777fc3ed9439a52a2206568 | [
"Unlicense"
] | null | null | null | main.py | Kubik-Kubik/cs-go-map-chooser | 82e2e27999d151a06777fc3ed9439a52a2206568 | [
"Unlicense"
] | null | null | null | main.py | Kubik-Kubik/cs-go-map-chooser | 82e2e27999d151a06777fc3ed9439a52a2206568 | [
"Unlicense"
] | null | null | null | from PyQt5 import QtWidgets, uic
import sys
from ui import Ui_Dialog
from PyQt5 import QtCore, QtGui, QtWidgets
import sys
from ui import Ui_Dialog
# create app
app = QtWidgets.QApplication(sys.argv)
# create dialog
Dialog = QtWidgets.QDialog()
ui = Ui_Dialog()
ui.setupUi(Dialog)
Dialog.setWindowTitle("Map Picker")
Dialog.setWindowIcon(QtGui.QIcon("logo.png"))
Dialog.show()
# hook logic
def reset():
ui.vertigo_img.setStyleSheet("QPushButton {background-image: url(images/vertigo.png);"
"border: none;}")
ui.inferno_img.setStyleSheet("QPushButton {background-image: url(images/inferno.png);"
"border: none;}")
ui.overpass_img.setStyleSheet("QPushButton {background-image: url(images/overpass.png);"
"border: none;}")
ui.cobble_img.setStyleSheet("QPushButton {background-image: url(images/cobble.png);"
"border: none;}")
ui.train_img.setStyleSheet("QPushButton {background-image: url(images/train.png);"
"border: none;}")
ui.dust_img.setStyleSheet("QPushButton {background-image: url(images/dust.png);"
"border: none;}")
ui.nuke_img.setStyleSheet("QPushButton {background-image: url(images/nuke.png);"
"border: none;}")
ui.realty_img.setStyleSheet("QPushButton {background-image: url(images/rialto.png);"
"border: none;}")
ui.lake_img.setStyleSheet("QPushButton {background-image: url(images/lake.png);"
"border: none;}")
def ban_map():
if ui.vertigo_btn.isChecked():
ui.vertigo_img.setStyleSheet("QPushButton {background-image: url(images/vertigo_ban.png);"
"border: none;}")
elif ui.inferno_btn.isChecked():
ui.inferno_img.setStyleSheet("QPushButton {background-image: url(images/inferno_ban.png);"
"border: none;}")
elif ui.overpass_btn.isChecked():
ui.overpass_img.setStyleSheet("QPushButton {background-image: url(images/overpass_ban.png);"
"border: none;}")
elif ui.cobble_btn.isChecked():
ui.cobble_img.setStyleSheet("QPushButton {background-image: url(images/cobble_ban.png);"
"border: none;}")
elif ui.train_btn.isChecked():
ui.train_img.setStyleSheet("QPushButton {background-image: url(images/train_ban.png);"
"border: none;}")
elif ui.dust_btn.isChecked():
ui.dust_img.setStyleSheet("QPushButton {background-image: url(images/dust_ban.png);"
"border: none;}")
elif ui.nuke_btn.isChecked():
ui.nuke_img.setStyleSheet("QPushButton {background-image: url(images/nuke_ban.png);"
"border: none;}")
elif ui.rialto_btn.isChecked():
ui.realty_img.setStyleSheet("QPushButton {background-image: url(images/rialto_ban.png);"
"border: none;}")
elif ui.lake_btn.isChecked():
ui.lake_img.setStyleSheet("QPushButton {background-image: url(images/lake_ban.png);"
"border: none;}")
def pick_map():
if ui.vertigo_btn.isChecked():
ui.vertigo_img.setStyleSheet("QPushButton {background-image: url(images/vertigo_pick.png);"
"border: none;}")
elif ui.inferno_btn.isChecked():
ui.inferno_img.setStyleSheet("QPushButton {background-image: url(images/inferno_pick.png);"
"border: none;}")
elif ui.overpass_btn.isChecked():
ui.overpass_img.setStyleSheet("QPushButton {background-image: url(images/overpass_pick.png);"
"border: none;}")
elif ui.cobble_btn.isChecked():
ui.cobble_img.setStyleSheet("QPushButton {background-image: url(images/cobble_pick.png);"
"border: none;}")
elif ui.train_btn.isChecked():
ui.train_img.setStyleSheet("QPushButton {background-image: url(images/train_pick.png);"
"border: none;}")
elif ui.dust_btn.isChecked():
ui.dust_img.setStyleSheet("QPushButton {background-image: url(images/dust_pick.png);"
"border: none;}")
elif ui.nuke_btn.isChecked():
ui.nuke_img.setStyleSheet("QPushButton {background-image: url(images/nuke_pick.png);"
"border: none;}")
elif ui.rialto_btn.isChecked():
ui.realty_img.setStyleSheet("QPushButton {background-image: url(images/rialto_pick.png);"
"border: none;}")
elif ui.lake_btn.isChecked():
ui.lake_img.setStyleSheet("QPushButton {background-image: url(images/lake_pick.png);"
"border: none;}")
ui.ban_btn.clicked.connect(ban_map)
ui.pick_btn.clicked.connect(pick_map)
ui.reset_btn.clicked.connect(reset)
# run main loop
sys.exit(app.exec_())
| 47.294643 | 102 | 0.58316 | 551 | 5,297 | 5.471869 | 0.108893 | 0.143284 | 0.241791 | 0.331343 | 0.824876 | 0.824876 | 0.806302 | 0.787065 | 0.787065 | 0.787065 | 0 | 0.000532 | 0.289975 | 5,297 | 111 | 103 | 47.720721 | 0.801117 | 0.009251 | 0 | 0.532609 | 0 | 0 | 0.375292 | 0.135035 | 0 | 0 | 0 | 0 | 0 | 1 | 0.032609 | false | 0.054348 | 0.065217 | 0 | 0.097826 | 0 | 0 | 0 | 0 | null | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 9 |
aa787d3354b46adc19835d621fb2b7e7853329aa | 3,701 | py | Python | events/migrations/0067_modify_several_fields_on_delete.py | jannetasa/linkedevents | 0d72c3266025b1bd71b3f328fd2f8a4b75e1b82b | [
"MIT"
] | 20 | 2015-05-28T16:02:00.000Z | 2021-07-14T06:36:19.000Z | events/migrations/0067_modify_several_fields_on_delete.py | jannetasa/linkedevents | 0d72c3266025b1bd71b3f328fd2f8a4b75e1b82b | [
"MIT"
] | 358 | 2015-02-04T09:07:19.000Z | 2022-03-28T12:10:22.000Z | events/migrations/0067_modify_several_fields_on_delete.py | jannetasa/linkedevents | 0d72c3266025b1bd71b3f328fd2f8a4b75e1b82b | [
"MIT"
] | 38 | 2015-02-23T13:16:02.000Z | 2021-12-13T07:48:23.000Z | # Generated by Django 2.2.9 on 2020-01-08 08:22
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('events', '0066_modify_datasource_owner_on_delete'),
]
operations = [
migrations.AlterField(
model_name='event',
name='created_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_event_created_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='event',
name='last_modified_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_event_modified_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='image',
name='created_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='EventImage_created_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='image',
name='last_modified_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='EventImage_last_modified_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='image',
name='license',
field=models.ForeignKey(default='cc_by', null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='images', to='events.License', verbose_name='License'),
),
migrations.AlterField(
model_name='keyword',
name='created_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_keyword_created_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='keyword',
name='last_modified_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_keyword_modified_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='keywordset',
name='created_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_keywordset_created_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='keywordset',
name='last_modified_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_keywordset_modified_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='place',
name='created_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_place_created_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='place',
name='last_modified_by',
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='events_place_modified_by', to=settings.AUTH_USER_MODEL),
),
migrations.AlterField(
model_name='place',
name='replaced_by',
field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='aliases', to='events.Place'),
),
]
| 48.697368 | 180 | 0.668738 | 444 | 3,701 | 5.304054 | 0.128378 | 0.047558 | 0.077282 | 0.121444 | 0.850955 | 0.850955 | 0.816136 | 0.816136 | 0.816136 | 0.816136 | 0 | 0.006531 | 0.213996 | 3,701 | 75 | 181 | 49.346667 | 0.803025 | 0.012159 | 0 | 0.666667 | 1 | 0 | 0.155172 | 0.078818 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.043478 | 0 | 0.086957 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
aaa2cbabba118d0912358463f786dc71ba0680b1 | 133 | py | Python | hypothalamus_seg/__init__.py | BBillot/hypothlamus_seg | a97bf109370ce758ad2ee457f12ba173ba7da818 | [
"Apache-2.0"
] | 21 | 2020-05-27T10:11:45.000Z | 2022-03-12T20:19:22.000Z | hypothalamus_seg/__init__.py | BBillot/hypothlamus_seg | a97bf109370ce758ad2ee457f12ba173ba7da818 | [
"Apache-2.0"
] | 5 | 2020-01-28T22:12:30.000Z | 2022-03-03T15:13:37.000Z | hypothalamus_seg/__init__.py | BBillot/hypothlamus_seg | a97bf109370ce758ad2ee457f12ba173ba7da818 | [
"Apache-2.0"
] | 5 | 2021-05-13T06:31:36.000Z | 2022-03-13T19:11:58.000Z | from . import augmentation_model
from . import model_inputs
from . import metrics_model
from . import predict
from . import training
| 22.166667 | 32 | 0.81203 | 18 | 133 | 5.833333 | 0.444444 | 0.47619 | 0.285714 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.150376 | 133 | 5 | 33 | 26.6 | 0.929204 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 7 |
2aee171d72ccbeed70c5a6947afde318c51db467 | 40 | py | Python | OMASS4/Opt_SNR_AR/__init__.py | DBernardes/OMASS4 | 30d2edc961463253cc120bc8ca1d74a0a73d922d | [
"MIT"
] | null | null | null | OMASS4/Opt_SNR_AR/__init__.py | DBernardes/OMASS4 | 30d2edc961463253cc120bc8ca1d74a0a73d922d | [
"MIT"
] | null | null | null | OMASS4/Opt_SNR_AR/__init__.py | DBernardes/OMASS4 | 30d2edc961463253cc120bc8ca1d74a0a73d922d | [
"MIT"
] | null | null | null | from .optimize_snr_ar import Opt_SNR_AR
| 20 | 39 | 0.875 | 8 | 40 | 3.875 | 0.75 | 0.322581 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.1 | 40 | 1 | 40 | 40 | 0.861111 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 1 | 0 | null | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 7 |
63077cb6d7c7f4ae64f4badcabcc599b41c0d809 | 57,404 | py | Python | mseg/label_preparation/dataset_update_records.py | khose/mseg-api | ae649c40918a5fc8a14a88a12b01b8032cc53492 | [
"CC-BY-4.0"
] | 213 | 2020-04-25T02:51:42.000Z | 2022-03-25T09:57:26.000Z | mseg/label_preparation/dataset_update_records.py | khose/mseg-api | ae649c40918a5fc8a14a88a12b01b8032cc53492 | [
"CC-BY-4.0"
] | 19 | 2020-04-28T05:47:50.000Z | 2022-03-18T03:05:07.000Z | mseg/label_preparation/dataset_update_records.py | khose/mseg-api | ae649c40918a5fc8a14a88a12b01b8032cc53492 | [
"CC-BY-4.0"
] | 35 | 2020-04-29T16:11:43.000Z | 2022-03-28T15:50:15.000Z | #!/usr/bin/python3
from mseg.label_preparation.relabeled_data_containers import DatasetClassUpdateRecord
""" COCO Panoptic"""
cocop_update_records = [
# cabinet-merged-classification,fence-guardrail-classification
# person_rider_classification,rug-merged,showercurtain_curtainother_classification
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'desk', 'cocop_cabinet-merged/cocopanoptic_desk_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'bookshelf', 'cocop_cabinet-merged/cocopanoptic_bookshelf_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'bookshelf', 'cocop_cabinet-merged/cocopanoptic_bookshelf_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'cabinet', 'cocop_cabinet-merged/cocopanoptic_cabinet_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'cabinet', 'cocop_cabinet-merged/cocopanoptic_cabinet_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'unlabeled', 'cocop_cabinet-merged/cocopanoptic_cabinetmerged_unlabel_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'unlabeled', 'cocop_cabinet-merged/cocopanoptic_cabinetmerged_unlabel_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'counter_other', 'cocop_cabinet-merged/cocopanoptic_counter_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'chest_of_drawers', 'cocop_cabinet-merged/cocopanoptic_dresser_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'chest_of_drawers', 'cocop_cabinet-merged/cocopanoptic_dresser_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'nightstand', 'cocop_cabinet-merged/cocopanoptic_nightstand_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'cabinet', 'nightstand', 'cocop_cabinet-merged/cocopanoptic_nightstand_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'cabinet', 'wardrobe', 'cocop_cabinet-merged/cocopanoptic_wardrobe_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'fence', 'fence', 'cocop_fence-guardrail/cocopanoptic_fence_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'fence', 'fence', 'cocop_fence-guardrail/cocopanoptic_fence_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'fence', 'guard_rail', 'cocop_fence-guardrail/cocopanoptic_guardrail_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'fence', 'guard_rail', 'cocop_fence-guardrail/cocopanoptic_guardrail_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'rug_floormat', 'unlabeled', 'cocop_rug-merged/cocopanoptic_rug-merged_unlabel-list_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'rug_floormat', 'unlabeled', 'cocop_rug-merged/cocopanoptic_rug-merged_unlabel-list_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'rug_floormat', 'floor', 'cocop_rug-merged/floor-carpet_80percent_confidence_cocopanoptic_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'rug_floormat', 'floor', 'cocop_rug-merged/floor-carpet_80percent_confidence_cocopanoptic_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'rug_floormat', 'rug_floormat', 'cocop_rug-merged/rug-floormat_80percent_confidence_cocopanoptic_train.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'rug_floormat', 'rug_floormat', 'cocop_rug-merged/rug-floormat_80percent_confidence_cocopanoptic_val.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'curtain_other', 'curtain_other', 'cocop_curtain/2019_08_16_coco_other-curtain_80percent_conf_valsplit.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'val', 'curtain_other', 'shower_curtain', 'cocop_curtain/2019_08_16_coco_showercurtain_80percent_conf_valsplit.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'curtain_other', 'curtain_other', 'cocop_curtain/2019_08_17_coco_other-curtain_train_80percent_conf.txt'),
DatasetClassUpdateRecord('cocopanoptic', 'train', 'curtain_other', 'shower_curtain', 'cocop_curtain/2019_08_17_coco_showercurtain_train_80percent_conf.txt'),
# cocop_cocop_chair (COCOP CHAIR-> UNIVERSAL CHAIR-OTHER )
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'armchair', 'cocop_chair/cocop_train_cocop_chair_to_armchair.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'basket', 'cocop_chair/cocop_train_cocop_chair_to_basket.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'bench', 'cocop_chair/cocop_train_cocop_chair_to_bench.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'chair_other', 'cocop_chair/cocop_train_cocop_chair_to_chair-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'unlabeled', 'cocop_chair/cocop_train_cocop_chair_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'ottoman', 'cocop_chair/cocop_train_cocop_chair_to_ottoman.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'seat', 'cocop_chair/cocop_train_cocop_chair_to_seat.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'slow_wheeled_object', 'cocop_chair/cocop_train_cocop_chair_to_slowwheeledobject.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'stool', 'cocop_chair/cocop_train_cocop_chair_to_stool.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'chair_other', 'swivel_chair', 'cocop_chair/cocop_train_cocop_chair_to_swivel-chair.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'armchair', 'cocop_chair/cocop_val_cocop_chair_to_armchair.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'chair_other', 'cocop_chair/cocop_val_cocop_chair_to_chair-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'unlabeled', 'cocop_chair/cocop_val_cocop_chair_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'ottoman', 'cocop_chair/cocop_val_cocop_chair_to_ottoman.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'seat', 'cocop_chair/cocop_val_cocop_chair_to_seat.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'slow_wheeled_object', 'cocop_chair/cocop_val_cocop_chair_to_slowwheeledobject.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'stool', 'cocop_chair/cocop_val_cocop_chair_to_stool.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'chair_other', 'swivel_chair', 'cocop_chair/cocop_val_cocop_chair_to_swivel-chair.txt'),
# cocop_cocop_counter (COCO COUNTER -> UNIVERSAL COUNTER-OTHER)
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'bathroom_counter', 'cocop_cocop_counter/cocop_train_cocop_counter_to_bathroom-counter.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'counter_other', 'cocop_cocop_counter/cocop_train_cocop_counter_to_counter-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'kitchen_island', 'cocop_cocop_counter/cocop_train_cocop_counter_to_kitchen-island.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'nightstand', 'cocop_cocop_counter/cocop_train_cocop_counter_to_nightstand.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'counter_other', 'unlabeled', 'cocop_cocop_counter/cocop_train_cocop_counter_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'counter_other', 'bathroom_counter', 'cocop_cocop_counter/cocop_val_cocop_counter_to_bathroom-counter.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'counter_other', 'counter_other', 'cocop_cocop_counter/cocop_val_cocop_counter_to_counter-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'counter_other', 'kitchen_island', 'cocop_cocop_counter/cocop_val_cocop_counter_to_kitchen-island.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'counter_other', 'unlabeled', 'cocop_cocop_counter/cocop_val_cocop_counter_to_None_of_these.txt'),
# cocop_cocop_diningtable (COCOP {table-merged, dining table} -> UNIVERSAL TABLE )
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'bathroom_counter', 'cocop_diningtable/cocop_train_cocop_diningtable_to_bathroom-counter.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'box', 'cocop_diningtable/cocop_train_cocop_diningtable_to_box.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'chest_of_drawers', 'cocop_diningtable/cocop_train_cocop_diningtable_to_chestofdrawers.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'counter_other', 'cocop_diningtable/cocop_train_cocop_diningtable_to_counter-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'desk', 'cocop_diningtable/cocop_train_cocop_diningtable_to_desk-and-table.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'desk', 'cocop_diningtable/cocop_train_cocop_diningtable_to_desk.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'kitchen_island', 'cocop_diningtable/cocop_train_cocop_diningtable_to_kitchen-island.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'nightstand', 'cocop_diningtable/cocop_train_cocop_diningtable_to_nightstand.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'unlabeled', 'cocop_diningtable/cocop_train_cocop_diningtable_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'table', 'cocop_diningtable/cocop_train_cocop_diningtable_to_table.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'box', 'cocop_diningtable/cocop_val_diningtable_to_box.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'counter_other', 'cocop_diningtable/cocop_val_diningtable_to_counterother.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'desk', 'cocop_diningtable/cocop_val_diningtable_to_desk.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'desk', 'cocop_diningtable/cocop_val_diningtable_to_deskandtable.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'kitchen_island', 'cocop_diningtable/cocop_val_diningtable_to_kitchenisland.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'table', 'cocop_diningtable/cocop_val_diningtable_to_table.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'unlabeled', 'cocop_diningtable/cocop_val_diningtable_to_unlabel.txt'),
# cocop_cocop_table-merged
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'base', 'cocop_table-merged/cocop_train_cocop_table-merged_to_base.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'bathroom_counter','cocop_table-merged/cocop_train_cocop_table-merged_to_bathroom-counter.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'cabinet', 'cocop_table-merged/cocop_train_cocop_table-merged_to_cabinet.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'chest_of_drawers','cocop_table-merged/cocop_train_cocop_table-merged_to_chestofdrawers.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'counter_other', 'cocop_table-merged/cocop_train_cocop_table-merged_to_counter-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'desk', 'cocop_table-merged/cocop_train_cocop_table-merged_to_desk.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'kitchen_island', 'cocop_table-merged/cocop_train_cocop_table-merged_to_kitchen-island.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'laptop', 'cocop_table-merged/cocop_train_cocop_table-merged_to_laptop.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'mountain_hill', 'cocop_table-merged/cocop_train_cocop_table-merged_to_mountainhill.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'nightstand', 'cocop_table-merged/cocop_train_cocop_table-merged_to_nightstand.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'unlabeled', 'cocop_table-merged/cocop_train_cocop_table-merged_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'pool_table', 'cocop_table-merged/cocop_train_cocop_table-merged_to_pooltable.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'stool', 'cocop_table-merged/cocop_train_cocop_table-merged_to_stool.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'desk', 'cocop_table-merged/cocop_train_cocop_table-merged_to_table-used-as-desk.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'table', 'table', 'cocop_table-merged/cocop_train_cocop_table-merged_to_table.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'bathroom_counter','cocop_table-merged/cocop_val_cocop_table-merged_to_bathroomcounter.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'counter_other', 'cocop_table-merged/cocop_val_cocop_table-merged_to_counterother.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'desk', 'cocop_table-merged/cocop_val_cocop_table-merged_to_desk.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'kitchen_island', 'cocop_table-merged/cocop_val_cocop_table-merged_to_kitchenisland.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'nightstand', 'cocop_table-merged/cocop_val_cocop_table-merged_to_nightstand.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'desk', 'cocop_table-merged/cocop_val_cocop_table-merged_to_table-used-as-desk.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'table', 'cocop_table-merged/cocop_val_cocop_table-merged_to_table.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'table', 'unlabeled', 'cocop_table-merged/cocop_val_cocop_table-merged_to_unlabel.txt'),
# cocop_cocop_light (COCO LIGHT-> UNIVERSAL LIGHT-OTHER)
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'chandelier', 'cocop_light/cocop_train_cocop_light_to_chandelier.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'lamp', 'cocop_light/cocop_train_cocop_light_to_lamp.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'light_other', 'cocop_light/cocop_train_cocop_light_to_light-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'unlabeled', 'cocop_light/cocop_train_cocop_light_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'sconce', 'cocop_light/cocop_train_cocop_light_to_sconce.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'light_other', 'streetlight', 'cocop_light/cocop_train_cocop_light_to_streetlight.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'chandelier', 'cocop_light/cocop_val_cocop_light_to_chandelier.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'lamp', 'cocop_light/cocop_val_cocop_light_to_lamp.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'light_other', 'cocop_light/cocop_val_cocop_light_to_light-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'unlabeled', 'cocop_light/cocop_val_cocop_light_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'sconce', 'cocop_light/cocop_val_cocop_light_to_sconce.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'light_other', 'streetlight', 'cocop_light/cocop_val_cocop_light_to_streetlight.txt'),
# cocopanoptic_cocop_motorcyclist_bicyclist (COCO PERSON -> UNIVERSAL PERSON-NON-RIDER)
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'bicyclist', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_bicyclist.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'motorcyclist', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_motorcyclist.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'unlabeled', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'person', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_person-non-rider.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'rider_other', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_train_cocop_motorcyclist_bicyclist_to_rider-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'bicyclist', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_val_cocop_motorcyclist_bicyclist_to_bicyclist.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'motorcyclist', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_val_cocop_motorcyclist_bicyclist_to_motorcyclist.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'person', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_val_cocop_motorcyclist_bicyclist_to_person-non-rider.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'rider_other', 'cocopanoptic_cocop_motorcyclist_bicyclist/cocopanoptic_val_cocop_motorcyclist_bicyclist_to_rider-other.txt'),
# cocopanoptic_cocop_nonrider
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'bicyclist', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_bicyclist.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'motorcyclist', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_motorcyclist.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'unlabeled', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'person', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'train', 'person', 'rider_other', 'cocopanoptic_cocop_nonrider/cocopanoptic_train_cocop_nonrider_to_rider-other.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'bicyclist', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_bicyclist.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'motorcyclist', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_motorcyclist.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'unlabeled', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'person', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('cocopanoptic-v1', 'val', 'person', 'rider_other', 'cocopanoptic_cocop_nonrider/cocopanoptic_val_cocop_nonrider_to_rider-other.txt'),
# Phase 3, cocop_waterother
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'fountain', 'cocop_waterother/cocop_train_cocop_waterother_to_fountain.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'unlabeled', 'cocop_waterother/cocop_train_cocop_waterother_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'playingfield', 'cocop_waterother/cocop_train_cocop_waterother_to_playingfield.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'river_lake', 'cocop_waterother/cocop_train_cocop_waterother_to_river-lake.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'sea', 'cocop_waterother/cocop_train_cocop_waterother_to_sea.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'swimming_pool', 'cocop_waterother/cocop_train_cocop_waterother_to_swimming-pool.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'terrain', 'cocop_waterother/cocop_train_cocop_waterother_to_terrain.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'wall', 'cocop_waterother/cocop_train_cocop_waterother_to_wall.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'water_other', 'cocop_waterother/cocop_train_cocop_waterother_to_water-other.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'waterfall', 'cocop_waterother/cocop_train_cocop_waterother_to_waterfall.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'water_other', 'window', 'cocop_waterother/cocop_train_cocop_waterother_to_window.txt'),
# Phase 3,cocop_runway_pavement
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocop','train', 'sidewalk_pavement', 'runway', 'cocop_runway_pavement/cocop_train_pavementmerged_to_airport_runway.txt'),
DatasetClassUpdateRecord('cocop','train', 'sidewalk_pavement', 'floor', 'cocop_runway_pavement/cocop_train_pavementmerged_to_floor.txt'),
DatasetClassUpdateRecord('cocop','train', 'sidewalk_pavement', 'sidewalk_pavement', 'cocop_runway_pavement/cocop_train_pavementmerged_to_pavementmerged.txt'),
DatasetClassUpdateRecord('cocop','train', 'sidewalk_pavement', 'pier_wharf', 'cocop_runway_pavement/cocop_train_pavementmerged_to_pierwharf.txt'),
DatasetClassUpdateRecord('cocop','val' , 'sidewalk_pavement', 'runway', 'cocop_runway_pavement/cocop_val_pavementmerged_to_airport_runway.txt'),
DatasetClassUpdateRecord('cocop','val' , 'sidewalk_pavement', 'sidewalk_pavement', 'cocop_runway_pavement/cocop_val_pavementmerged_to_pavementmerged.txt'),
# Phase 3,cocop_runway_road
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocop', 'train', 'road', 'runway', 'cocop_runway_road/cocop_train_road_to_airportrunway.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'road', 'sidewalk_pavement','cocop_runway_road/cocop_train_road_to_pavementsidewalk.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'road', 'road', 'cocop_runway_road/cocop_train_road_to_road.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'road', 'unlabeled', 'cocop_runway_road/cocop_train_road_to_unlabel.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'road', 'runway', 'cocop_runway_road/cocop_val_road_to_airportrunway.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'road', 'sidewalk_pavement','cocop_runway_road/cocop_val_road_to_pavementsidewalk.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'road', 'road', 'cocop_runway_road/cocop_val_road_to_road.txt'),
# Phase 3,cocop_cocop_platform
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocop', 'train', 'platform', 'bridge', 'cocop_platform/cocop_train_cocop_platform_to_bridge.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'platform', 'unlabeled', 'cocop_platform/cocop_train_cocop_platform_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'platform', 'pier_wharf', 'cocop_platform/cocop_train_cocop_platform_to_pier-wharf.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'platform', 'platform', 'cocop_platform/cocop_train_cocop_platform_to_platform.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'platform', 'building', 'cocop_platform/cocop_val_cocop_platform_to_grandstand.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'platform', 'unlabeled', 'cocop_platform/cocop_val_cocop_platform_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'platform', 'pier_wharf', 'cocop_platform/cocop_val_cocop_platform_to_pier-wharf.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'platform', 'platform', 'cocop_platform/cocop_val_cocop_platform_to_platform.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'platform', 'playingfield', 'cocop_platform/cocop_val_cocop_platform_to_playingfield.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'platform', 'stage', 'cocop_platform/cocop_val_cocop_platform_to_stage.txt'),
# Phase 3, cocop_cocop_bridge
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'bridge', 'cocop_bridge/cocop_train_cocop_bridge_to_bridge.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'building', 'cocop_bridge/cocop_train_cocop_bridge_to_building.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'building', 'cocop_bridge/cocop_train_cocop_bridge_to_grandstand.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'building', 'cocop_bridge/cocop_train_cocop_bridge_to_house.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'unlabeled', 'cocop_bridge/cocop_train_cocop_bridge_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'pier_wharf', 'cocop_bridge/cocop_train_cocop_bridge_to_pier-wharf.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'platform', 'cocop_bridge/cocop_train_cocop_bridge_to_platform.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'runway', 'cocop_bridge/cocop_train_cocop_bridge_to_runway.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'bridge', 'vegetation', 'cocop_bridge/cocop_train_cocop_bridge_to_vegetation.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'bridge', 'cocop_bridge/cocop_val_cocop_bridge_to_bridge.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'building', 'cocop_bridge/cocop_val_cocop_bridge_to_building.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'building', 'cocop_bridge/cocop_val_cocop_bridge_to_house.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'unlabeled', 'cocop_bridge/cocop_val_cocop_bridge_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'bridge', 'pier_wharf', 'cocop_bridge/cocop_val_cocop_bridge_to_pier-wharf.txt'),
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'awning', 'cocop_tent/cocop_train_cocop_tent_to_awning.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'base', 'cocop_tent/cocop_train_cocop_tent_to_base.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'building', 'cocop_tent/cocop_train_cocop_tent_to_building.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'fence', 'cocop_tent/cocop_train_cocop_tent_to_fence.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'unlabeled', 'cocop_tent/cocop_train_cocop_tent_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'billboard', 'cocop_tent/cocop_train_cocop_tent_to_signboard.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'tent', 'cocop_tent/cocop_train_cocop_tent_to_tent.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'umbrella', 'cocop_tent/cocop_train_cocop_tent_to_umbrella.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'vegetation', 'cocop_tent/cocop_train_cocop_tent_to_vegetation.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'wall', 'cocop_tent/cocop_train_cocop_tent_to_wall.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'window', 'cocop_tent/cocop_train_cocop_tent_to_window.txt'),
DatasetClassUpdateRecord('cocop', 'train', 'tent', 'kite', 'cocop_tent/cocop_train_cocop_tent_to_windsurfingkite.txt'),
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cocop', 'val', 'tent', 'awning', 'cocop_tent/cocop_val_cocop_tent_to_awning.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'tent', 'building', 'cocop_tent/cocop_val_cocop_tent_to_building.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'tent', 'unlabeled', 'cocop_tent/cocop_val_cocop_tent_to_None_of_these.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'tent', 'tent', 'cocop_tent/cocop_val_cocop_tent_to_tent.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'tent', 'umbrella', 'cocop_tent/cocop_val_cocop_tent_to_umbrella.txt'),
DatasetClassUpdateRecord('cocop', 'val', 'tent', 'window', 'cocop_tent/cocop_val_cocop_tent_to_window.txt')
]
""" ADE20K """
ade20k_update_records = [
# Phase1 - fence guardrail, person-rider, shower-curtain, curtain-other, snow-mountain-hill-ade20k
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('ade20k-v1', 'train', 'fence', 'fence', 'ade20k_fence_guardrail/ade20k_fence_train.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'fence', 'fence', 'ade20k_fence_guardrail/ade20k_fence_val.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'fence', 'guard_rail', 'ade20k_fence_guardrail/ade20k_guardrail_train.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'fence', 'guard_rail', 'ade20k_fence_guardrail/ade20k_guardrail_val.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'curtain_other', 'curtain_other', 'ade20k_curtain/2019_10_31_ade20k_other-curtain_valsplit.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'curtain_other', 'shower_curtain', 'ade20k_curtain/2019_10_31_ade20k_showercurtain_valsplit.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'curtain_other', 'curtain_other', 'ade20k_curtain/2019_11_1_ade20k_other-curtain_trainsplit.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'curtain_other', 'shower_curtain', 'ade20k_curtain/2019_11_1_ade20k_showercurtain_trainsplit.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'mountain_hill', 'mountain_hill', 'ade20k_snow-mountain-hill/ade20k_mountainhill_train.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'mountain_hill', 'mountain_hill', 'ade20k_snow-mountain-hill/ade20k_mountainhill_val.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'mountain_hill', 'snow', 'ade20k_snow-mountain-hill/ade20k_snow_train.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'mountain_hill', 'snow', 'ade20k_snow-mountain-hill/ade20k_snow_val.txt'),
# ADE20K_ade20k_animal
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'animal_other', 'ade20k_animal/ADE20K_train_ade20k_animal_to_animal-other.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'bear', 'ade20k_animal/ADE20K_train_ade20k_animal_to_bear.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'bird', 'ade20k_animal/ADE20K_train_ade20k_animal_to_bird.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'cat', 'ade20k_animal/ADE20K_train_ade20k_animal_to_cat.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'cow', 'ade20k_animal/ADE20K_train_ade20k_animal_to_cow.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'dog', 'ade20k_animal/ADE20K_train_ade20k_animal_to_dog.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'elephant', 'ade20k_animal/ADE20K_train_ade20k_animal_to_elephant.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'giraffe', 'ade20k_animal/ADE20K_train_ade20k_animal_to_giraffe.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'horse', 'ade20k_animal/ADE20K_train_ade20k_animal_to_horse.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'unlabeled', 'ade20k_animal/ADE20K_train_ade20k_animal_to_None_of_these.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other','plaything_other','ade20k_animal/ADE20K_train_ade20k_animal_to_plaything.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'sheep', 'ade20k_animal/ADE20K_train_ade20k_animal_to_sheep.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'animal_other', 'zebra', 'ade20k_animal/ADE20K_train_ade20k_animal_to_zebra.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'animal_other', 'ade20k_animal/ADE20K_val_ade20k_animal_to_animal-other.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'bear', 'ade20k_animal/ADE20K_val_ade20k_animal_to_bear.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'bird', 'ade20k_animal/ADE20K_val_ade20k_animal_to_bird.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'cow', 'ade20k_animal/ADE20K_val_ade20k_animal_to_cow.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'elephant', 'ade20k_animal/ADE20K_val_ade20k_animal_to_elephant.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'horse', 'ade20k_animal/ADE20K_val_ade20k_animal_to_horse.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'animal_other', 'sheep', 'ade20k_animal/ADE20K_val_ade20k_animal_to_sheep.txt'),
# ade20k_ade20k_chest_of_drawers
DatasetClassUpdateRecord('ade20k-v1', 'train', 'chest_of_drawers', 'cabinet', 'ade20k_chest_of_drawers/ade20k_train_ade20k_chest_of_drawers_to_cabinet.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'chest_of_drawers', 'chest_of_drawers', 'ade20k_chest_of_drawers/ade20k_train_ade20k_chest_of_drawers_to_chest-of-drawers.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'chest_of_drawers', 'nightstand', 'ade20k_chest_of_drawers/ade20k_train_ade20k_chest_of_drawers_to_nightstand.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'chest_of_drawers', 'unlabeled', 'ade20k_chest_of_drawers/ade20k_train_ade20k_chest_of_drawers_to_None_of_these.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'chest_of_drawers', 'cabinet', 'ade20k_chest_of_drawers/ade20k_val_ade20k_chest_of_drawers_to_cabinet.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'chest_of_drawers', 'chest_of_drawers', 'ade20k_chest_of_drawers/ade20k_val_ade20k_chest_of_drawers_to_chest-of-drawers.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'chest_of_drawers', 'desk', 'ade20k_chest_of_drawers/ade20k_val_ade20k_chest_of_drawers_to_desk.txt'),
# TODO
# ADE20K_ade20k_food
DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'cake', 'ade20k_food/ADE20K_train_ade20k_food_to_cake.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'food_other', 'ade20k_food/ADE20K_train_ade20k_food_to_foodother.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'fruit_other', 'ade20k_food/ADE20K_train_ade20k_food_to_fruitother.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'pizza', 'ade20k_food/ADE20K_train_ade20k_food_to_pizza.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'sandwich', 'ade20k_food/ADE20K_train_ade20k_food_to_sandwich.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'unlabeled', 'ade20k_food/ADE20K_train_ade20k_food_to_unlabel.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'food_other', 'vegetation', 'ade20k_food/ADE20K_train_ade20k_food_to_vegetation.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'food_other', 'broccoli', 'ade20k_food/ADE20K_val_ade20k_food_to_broccoli.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'food_other', 'food_other', 'ade20k_food/ADE20K_val_ade20k_food_to_food-other.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'food_other', 'fruit_other', 'ade20k_food/ADE20K_val_ade20k_food_to_fruit-other.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'food_other', 'pizza', 'ade20k_food/ADE20K_val_ade20k_food_to_pizza.txt'),
# ADE20K_ade20k_glass (ADE20K GLASS -> UNIVERSAL WINE-GLASS)
DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'bottle', 'ade20k_glass/ADE20K_train_ade20k_glass_to_bottle.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'cup', 'ade20k_glass/ADE20K_train_ade20k_glass_to_cup.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'mirror', 'ade20k_glass/ADE20K_train_ade20k_glass_to_mirror.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'unlabeled', 'ade20k_glass/ADE20K_train_ade20k_glass_to_None_of_these.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'window', 'ade20k_glass/ADE20K_train_ade20k_glass_to_window.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'wine_glass', 'wine_glass', 'ade20k_glass/ADE20K_train_ade20k_glass_to_wine_glass.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'wine_glass', 'cup', 'ade20k_glass/ADE20K_val_ade20k_glass_to_cup.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'wine_glass', 'unlabeled', 'ade20k_glass/ADE20K_val_ade20k_glass_to_unlabel.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'wine_glass', 'window', 'ade20k_glass/ADE20K_val_ade20k_glass_to_window.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'wine_glass', 'wine_glass', 'ade20k_glass/ADE20K_val_ade20k_glass_to_wine_glass.txt'),
# ADE20K_ade20k_motorcyclist_bicyclist (ADE20K PERSON -> UNIVERSAL PERSON-NON-RIDER)
DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'bicyclist', 'ade20k_motorcyclist_bicyclist/ADE20K_train_ade20k_motorcyclist_bicyclist_to_bicyclist.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'motorcyclist', 'ade20k_motorcyclist_bicyclist/ADE20K_train_ade20k_motorcyclist_bicyclist_to_motorcyclist.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'rider_other', 'ade20k_motorcyclist_bicyclist/ADE20K_train_ade20k_motorcyclist_bicyclist_to_rider-other.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'person', 'bicyclist', 'ade20k_motorcyclist_bicyclist/ADE20K_val_ade20k_motorcyclist_bicyclist_to_bicyclist.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'person', 'motorcyclist', 'ade20k_motorcyclist_bicyclist/ADE20K_val_ade20k_motorcyclist_bicyclist_to_motorcyclist.txt'),
# ade20k_ade20k_nonrider (ADE20K PERSON -> UNIVERSAL PERSON-NON-RIDER)
DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'bicyclist', 'ade20k_nonrider/ade20k_train_ade20k_nonrider_to_bicyclist.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'person', 'person', 'ade20k_nonrider/ade20k_train_ade20k_nonrider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'person', 'person', 'ade20k_nonrider/ade20k_val_ade20k_nonrider_to_person-non-rider.txt'),
#ade20k_ade20k_table (ADE20K TABLE -> UNIVERSAL TABLE)
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'bathroom_counter', 'ade20k_table/ade20k_train_ade20k_table_to_bathroom-counter.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'cabinet', 'ade20k_table/ade20k_train_ade20k_table_to_cabinet.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'counter_other', 'ade20k_table/ade20k_train_ade20k_table_to_counter-other.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'desk', 'ade20k_table/ade20k_train_ade20k_table_to_desk-and-table.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'desk', 'ade20k_table/ade20k_train_ade20k_table_to_desk.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'kitchen_island', 'ade20k_table/ade20k_train_ade20k_table_to_kitchen-island.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'nightstand', 'ade20k_table/ade20k_train_ade20k_table_to_nightstand.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'unlabeled', 'ade20k_table/ade20k_train_ade20k_table_to_None_of_these.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'train', 'table', 'table', 'ade20k_table/ade20k_train_ade20k_table_to_table.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'counter_other', 'ade20k_table/ade20k_val_ade20k_table_to_counter-other.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'desk', 'ade20k_table/ade20k_val_ade20k_table_to_desk-and-table.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'desk', 'ade20k_table/ade20k_val_ade20k_table_to_desk.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'kitchen_island', 'ade20k_table/ade20k_val_ade20k_table_to_kitchen-island.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'nightstand', 'ade20k_table/ade20k_val_ade20k_table_to_nightstand.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'unlabeled', 'ade20k_table/ade20k_val_ade20k_table_to_None_of_these.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'stool', 'ade20k_table/ade20k_val_ade20k_table_to_stool.txt'),
DatasetClassUpdateRecord('ade20k-v1', 'val', 'table', 'table', 'ade20k_table/ade20k_val_ade20k_table_to_table.txt'),
# Phase 3, ade20k_plaything
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('ade20k', 'train', 'plaything_other', 'plaything_other','ade20k_plaything/ade20k_train_ade20k_plaything_to_plaything-other.txt'),
DatasetClassUpdateRecord('ade20k', 'train', 'plaything_other', 'teddy_bear', 'ade20k_plaything/ade20k_train_ade20k_plaything_to_teddy-bear.txt'),
DatasetClassUpdateRecord('ade20k', 'val', 'plaything_other', 'plaything_other', 'ade20k_plaything/ade20k_val_ade20k_plaything_to_plaything-other.txt'),
DatasetClassUpdateRecord('ade20k', 'val', 'plaything_other', 'teddy_bear', 'ade20k_plaything/ade20k_val_ade20k_plaything_to_teddy-bear.txt')
]
cityscapes_update_records = [
# cityscapes_rider, Phase 2 (CITYSCAPES RIDER -> UNVERISAL RIDER-OTHER)
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'bicycle', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_bicycle.txt'),
DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'bicyclist', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_bicyclist.txt'),
DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'motorcyclist', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_motorcyclist.txt'),
DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'unlabeled', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_None_of_these.txt'),
DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'person', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('cityscapes', 'train', 'rider_other', 'rider_other', 'cityscapes_rider/cityscapes_train_cityscapes_rider_to_rider-other.txt'),
DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'bicycle', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_bicycle.txt'),
DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'bicyclist', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_bicyclist.txt'),
DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'motorcyclist', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_motorcyclist.txt'),
DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'unlabeled', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_None_of_these.txt'),
DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'person', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('cityscapes', 'val', 'rider_other', 'rider_other', 'cityscapes_rider/cityscapes_val_cityscapes_rider_to_rider-other.txt')
]
bdd_update_records = [
# bdd person, Phase 2 (BDD PERSON -> UNIVERSAL PERSON-NONRIDER)
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('bdd', 'train', 'person', 'bicyclist', 'bdd_person/bdd_train_bdd_person_to_bicyclist.txt'),
DatasetClassUpdateRecord('bdd', 'train', 'person', 'motorcyclist', 'bdd_person/bdd_train_bdd_person_to_motorcyclist.txt'),
DatasetClassUpdateRecord('bdd', 'train', 'person', 'unlabeled', 'bdd_person/bdd_train_bdd_person_to_None_of_these.txt'),
DatasetClassUpdateRecord('bdd', 'train', 'person', 'person', 'bdd_person/bdd_train_bdd_person_to_person-non-rider.txt'),
DatasetClassUpdateRecord('bdd', 'val', 'person', 'bicyclist', 'bdd_person/bdd_val_bdd_person_to_bicyclist.txt'),
DatasetClassUpdateRecord('bdd', 'val', 'person', 'person', 'bdd_person/bdd_val_bdd_person_to_person-non-rider.txt'),
DatasetClassUpdateRecord('bdd', 'val', 'person', 'unlabeled', 'bdd_person/bdd_val_bdd_person_to_unlabel.txt'),
# bdd_rider, Phase 2 (BDD RIDER -> UNIVERSAL RIDER-OTHER))
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'bicyclist', 'bdd_rider/bdd_train_bdd_rider_to_bicyclist.txt'),
DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'motorcyclist', 'bdd_rider/bdd_train_bdd_rider_to_motorcyclist.txt'),
DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'unlabeled', 'bdd_rider/bdd_train_bdd_rider_to_None_of_these.txt'),
DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'person', 'bdd_rider/bdd_train_bdd_rider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('bdd', 'train', 'rider_other', 'rider_other', 'bdd_rider/bdd_train_bdd_rider_to_rider-other.txt'),
DatasetClassUpdateRecord('bdd', 'val', 'rider_other', 'bicyclist', 'bdd_rider/bdd_val_bdd_rider_to_bicyclist.txt'),
DatasetClassUpdateRecord('bdd', 'val', 'rider_other', 'motorcyclist', 'bdd_rider/bdd_val_bdd_rider_to_motorcyclist.txt'),
DatasetClassUpdateRecord('bdd', 'val', 'rider_other', 'person', 'bdd_rider/bdd_val_bdd_rider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('bdd', 'val', 'rider_other', 'rider_other', 'bdd_rider/bdd_val_bdd_rider_to_rider-other.txt')
]
idd_update_records = [
# idd_idd_rider, Phase 2 (IDD-NEW RIDER -> UNIVERSAL RIDER-OTHER)
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'backpack', 'idd_rider/idd_train_idd_rider_to_backpack.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'bag', 'idd_rider/idd_train_idd_rider_to_bag.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'bicycle', 'idd_rider/idd_train_idd_rider_to_bicycle.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'bicyclist', 'idd_rider/idd_train_idd_rider_to_bicyclist.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'box', 'idd_rider/idd_train_idd_rider_to_box.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'motorcycle', 'idd_rider/idd_train_idd_rider_to_motorcycle.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'motorcyclist', 'idd_rider/idd_train_idd_rider_to_motorcyclist.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'unlabeled', 'idd_rider/idd_train_idd_rider_to_None_of_these.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'person', 'idd_rider/idd_train_idd_rider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'pole', 'idd_rider/idd_train_idd_rider_to_pole.txt'),
DatasetClassUpdateRecord('idd', 'train', 'rider_other', 'rider_other', 'idd_rider/idd_train_idd_rider_to_rider-other.txt'),
DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'bicyclist', 'idd_rider/idd_val_idd_rider_to_bicyclist.txt'),
DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'motorcyclist', 'idd_rider/idd_val_idd_rider_to_motorcyclist.txt'),
DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'unlabeled', 'idd_rider/idd_val_idd_rider_to_None_of_these.txt'),
DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'person', 'idd_rider/idd_val_idd_rider_to_person-non-rider.txt'),
DatasetClassUpdateRecord('idd', 'val', 'rider_other', 'rider_other', 'idd_rider/idd_val_idd_rider_to_rider-other.txt')
]
# Note: we refer to the `test` split as `val`, since `val` not easily defined/found.
sunrgbd_update_records = [
# sunrgbd_chair
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'armchair', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_armchair.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'bench', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_bench.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'chair_other', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_chairother.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'seat', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_seat.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'couch', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_sofa.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'stool', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_stool.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'swivel_chair', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_swivelchair.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'chair_other', 'unlabeled', 'sunrgbd_chair/sunrgbd_test_sunrgbd_chair_to_unlabel.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'armchair', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_armchair.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'bench', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_bench.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'cabinet', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_cabinet.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'chair_other', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_chair-other.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'door', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_door.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'unlabeled', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_None_of_these.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'ottoman', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_ottoman.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'seat', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_seat.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'couch', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_sofa.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'stool', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_stool.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'swivel_chair', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_swivel-chair.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'table', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_table.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','chair_other', 'wall', 'sunrgbd_chair/sunrgbd_train_sunrgbd_chair_to_wall.txt'),
# sunrgbd_counter
DatasetClassUpdateRecord('sunrgbd', 'val', 'counter_other', 'bathroom_counter','sunrgbd_counter/sunrgbd_test_sunrgbd_counter_to_bathroomcounter.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'counter_other', 'counter_other', 'sunrgbd_counter/sunrgbd_test_sunrgbd_counter_to_counter-other.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'counter_other', 'unlabeled', 'sunrgbd_counter/sunrgbd_test_sunrgbd_counter_to_unlabel.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'bathroom_counter','sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_bathroom-counter.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'cabinet', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_cabinetmerged.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'counter_other', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_counter-other.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'desk', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_desk.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'kitchen_island', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_kitchen-island.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','counter_other', 'unlabeled', 'sunrgbd_counter/sunrgbd_train_sunrgbd_counter_to_None_of_these.txt'),
# sunrgbd_lamp (TODO: map it to "lamp" inside of the universal tax.)
DatasetClassUpdateRecord('sunrgbd', 'val', 'light_other', 'lamp', 'sunrgbd_lamp/sunrgbd_test_sunrgbd_lamp_to_lamp.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'light_other', 'sconce', 'sunrgbd_lamp/sunrgbd_test_sunrgbd_lamp_to_sconce.txt'),
DatasetClassUpdateRecord('sunrgbd', 'val', 'light_other', 'unlabeled', 'sunrgbd_lamp/sunrgbd_test_sunrgbd_lamp_to_unlabel.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'chandelier', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_chandelier.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'lamp', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_lamp.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'light_other', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_light-other.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'unlabeled', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_None_of_these.txt'),
DatasetClassUpdateRecord('sunrgbd', 'train','light_other', 'sconce', 'sunrgbd_lamp/sunrgbd_train_sunrgbd_lamp_to_sconce.txt')
]
mapillary_update_records = [
# mapillarypublic_water
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'fountain', 'mapillarypublic_water/mapillarypublic_train_water_to_fountain.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'unlabeled', 'mapillarypublic_water/mapillarypublic_train_water_to_None_of_these.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'river_lake', 'mapillarypublic_water/mapillarypublic_train_water_to_river-lake.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'sea', 'mapillarypublic_water/mapillarypublic_train_water_to_sea.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'water_other', 'water_other','mapillarypublic_water/mapillarypublic_train_water_to_water-other.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'val', 'water_other', 'fountain', 'mapillarypublic_water/mapillarypublic_val_water_to_fountain.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'val', 'water_other', 'river_lake', 'mapillarypublic_water/mapillarypublic_val_water_to_river-lake.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'val', 'water_other', 'sea', 'mapillarypublic_water/mapillarypublic_val_water_to_sea.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'val', 'water_other', 'water_other', 'mapillarypublic_water/mapillarypublic_val_water_to_water-other.txt'),
# mapillarypublic_groundanimal
# dataset # split # orig u name # new u name # path to txt file
DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'animal_other', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_animalother.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'bird', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_bird.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'cat', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_cat.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'cow', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_cow.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'dog', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_dog.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'horse', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_horse.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'train', 'dog', 'unlabeled', 'mapillarypublic_groundanimal/mapillaryacademic_train_groundanimal_to_none.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'val', 'dog', 'bird', 'mapillarypublic_groundanimal/mapillaryacademic_val_groundanimal_to_bird.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'val', 'dog', 'dog', 'mapillarypublic_groundanimal/mapillaryacademic_val_groundanimal_to_dog.txt'),
DatasetClassUpdateRecord('mapillary-public65', 'val', 'dog', 'unlabeled', 'mapillarypublic_groundanimal/mapillaryacademic_val_groundanimal_to_none.txt')
]
| 105.911439 | 195 | 0.785433 | 6,608 | 57,404 | 6.407234 | 0.034201 | 0.221923 | 0.107773 | 0.088122 | 0.939394 | 0.897494 | 0.739885 | 0.420393 | 0.295236 | 0.194147 | 0 | 0.019679 | 0.076702 | 57,404 | 541 | 196 | 106.107209 | 0.779155 | 0.050763 | 0 | 0 | 0 | 0 | 0.638556 | 0.418052 | 0 | 0 | 0 | 0.001848 | 0 | 1 | 0 | false | 0 | 0.002525 | 0 | 0.002525 | 0 | 0 | 0 | 0 | null | 1 | 0 | 0 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
2d6b442d289e60626f3caad22d37bc5cb6431b4b | 9,758 | py | Python | kvlnode/core/KVLSegment.py | paoloriccardi/key-value-log | bc08d26ec614d9570eef5ca567dc69cc1f9a6ef7 | [
"BSD-3-Clause"
] | null | null | null | kvlnode/core/KVLSegment.py | paoloriccardi/key-value-log | bc08d26ec614d9570eef5ca567dc69cc1f9a6ef7 | [
"BSD-3-Clause"
] | null | null | null | kvlnode/core/KVLSegment.py | paoloriccardi/key-value-log | bc08d26ec614d9570eef5ca567dc69cc1f9a6ef7 | [
"BSD-3-Clause"
] | null | null | null | import sys
import os
import json
from datetime import datetime
#Segment with JSON as values delimited by {}
class KVLSegmentJSON():
def __init__(self,filename):
self.homedir = os.path.join(os.path.abspath(os.path.dirname(__file__)), "files/")
self.filename = filename
self.filepath = self.homedir + self.filename
try:
self.file = open(self.filepath,"a+")
except OSError:
print ("Could not open file" + self.filepath + "\n")
sys.exit()
def appendKeyValue(self, key, value):
if not self.checkValueFormat(value):
print (value)
print ("json format error in input")
exit()
#stringValue = json.dumps(value)
elementString = str(key) + ":" + value + ";"
rwpointer = self.file.tell()
if rwpointer > 1:
#takes into account element separator ;
rwpointer = rwpointer
try:
self.file.write(elementString)
except OSError:
print ("Could not append Key:Value \n")
return -1
return rwpointer
def checkValueFormat(self,value):
try:
json.loads(value)
except ValueError as e:
print("Error in JSON format value " + str(e))
return False
return True
def retrieveElement(self,offset):
self.file.seek(offset)
elementString = ""
while True:
character = self.file.read(1)
if not character:
break
if character == "}":
elementString = elementString + character
break
elementString = elementString + character
return elementString
def retrieveValue(self,offset):
self.file.seek(offset)
valueString = ""
scanningValue = False
while True:
character = self.file.read(1)
if not character:
break
if not scanningValue and character == "{":
valueString = character
scanningValue = True
elif not scanningValue and character != "{":
pass
elif scanningValue and character == "}":
valueString = valueString + character
break
else:
valueString = valueString + character
return valueString
def createIndex(self):
#scan one char at the time the segment file from beginning to end which is very inefficient
#this method should be used only when a consistent/updated version of index is not available (e.g. from an existing bucket)
self.file.seek(0)
index = {}
prevchar = self.file.read(1)
if not prevchar:
return index
#non trivial case
self.file.seek(0)
key = ""
scanningKey = True
prevchar = ""
rwindex = 0
offset = 0
for char in self.file.read():
if scanningKey and char == ":":
scanningKey = False
index[key]=offset
key = ""
elif scanningKey and char != ":":
key = key + char
elif not scanningKey and prevchar == "}" and char == ";":
scanningKey = True
offset = rwindex +1
prevchar = char
rwindex = rwindex + 1
return index
def getTombstoneValue(self):
return "{}"
def flush(self):
#close segment file
self.file.close()
def attachNewFile(self,newFilename):
self.filename = newFilename
try:
self.filepath = self.homedir + self.filename
self.file = open(self.filepath, "a+")
except OSError:
print ("Could not open file" + newFilename + "\n")
sys.exit()
def shrinkToNewFile(self):
inmemoryKV = self.inMemoryKeyValue()
newFilename = self.generateFilename()
self.flush()
self.attachNewFile(newFilename)
for key,value in inmemoryKV.items():
self.appendKeyValue(key,value)
newIndex = self.createIndex()
return newIndex
def inMemoryKeyValue(self):
KVDict = {}
index = self.createIndex()
for key,offset in index.items():
value = self.retrieveValue(offset)
KVDict[key]=value
return KVDict
def initializeSegment(self,kvdict):
#careful when using this method straight from the API initialize, this method expects a dictionary
#in case of nested json in request from the API the situation should be handled before calling
#appendKeyValue (where value should be a valid JSON)
newFilename = self.generateFilename()
self.file.flush()
self.attachNewFile(newFilename)
for key,value in kvdict.items():
self.appendKeyValue(key,value)
newIndex = self.createIndex()
return newIndex
def generateFilename(self):
now = datetime.now()
prefix = now.strftime('%f')
newFilename = prefix + self.filename
return newFilename
#Segment with value equal to simple value, key:values separated by ;
class KVLSegmentSimpleValue():
def __init__(self,filename):
self.homedir = os.path.join(os.path.abspath(os.path.dirname(__file__)), "files/")
self.filename = filename
self.filepath = self.homedir + self.filename
try:
self.file = open(self.filepath,"a+")
except OSError:
print ("Could not open file" + self.filepath + "\n")
sys.exit()
def appendKeyValue(self, key, value):
elementString = str(key) + ":" + str(value) + ";"
rwpointer = self.file.tell()
if rwpointer > 1:
#takes into account element separator ;
rwpointer = rwpointer
try:
self.file.write(elementString)
except OSError:
print ("Could not append Key:Value \n")
return -1
return rwpointer
def retrieveElement(self,offset):
self.file.seek(offset)
elementString = ""
while True:
character = self.file.read(1)
if not character:
break
if character == ";":
elementString = elementString
break
elementString = elementString + character
return elementString
def retrieveValue(self,offset):
self.file.seek(offset)
valueString = ""
scanningValue = False
while True:
character = self.file.read(1)
if not character:
break
if not scanningValue and character == ":":
scanningValue = True
elif not scanningValue and character != ":":
pass
elif scanningValue and character == ";":
valueString = valueString
break
else:
valueString = valueString + character
return valueString
def createIndex(self):
#scan one char at the time the segment file from beginning to end which is very inefficient
#this method should be used only when a consistent/updated version of index is not available (e.g. from an existing bucket)
self.file.seek(0)
index = {}
if not self.file.read(1):
return index
#non trivial case
self.file.seek(0)
key = ""
scanningKey = True
rwindex = 0
offset = 0
for char in self.file.read():
if scanningKey and char == ":":
scanningKey = False
index[key]=offset
key = ""
elif scanningKey and char != ":":
key = key + char
elif not scanningKey and char != ";":
pass
elif not scanningKey and char == ";":
scanningKey = True
offset = rwindex +1
rwindex = rwindex + 1
return index
def getTombstoneValue(self):
return ""
def checkValueFormat(self,value):
return True
def flush(self):
#close segment file
self.file.close()
def attachNewFile(self,newFilename):
self.filename = newFilename
try:
self.filepath = self.homedir + self.filename
self.file = open(self.filepath, "a+")
except OSError:
print ("Could not open file" + newFilename + "\n")
sys.exit()
def shrinkToNewFile(self):
inmemoryKV = self.inMemoryKeyValue()
newFilename = self.generateFilename()
self.flush()
self.attachNewFile(newFilename)
for key,value in inmemoryKV.items():
self.appendKeyValue(key,value)
newIndex = self.createIndex()
return newIndex
def inMemoryKeyValue(self):
KVDict = {}
index = self.createIndex()
for key,offset in index.items():
value = self.retrieveValue(offset)
KVDict[key]=value
return KVDict
def initializeSegment(self,kvdict):
newFilename = self.generateFilename()
self.file.flush()
self.attachNewFile(newFilename)
for key,value in kvdict.items():
self.appendKeyValue(key,value)
newIndex = self.createIndex()
return newIndex
def generateFilename(self):
now = datetime.now()
prefix = now.strftime('%f')
newFilename = prefix + self.filename
return newFilename
| 31.993443 | 131 | 0.552572 | 958 | 9,758 | 5.611691 | 0.160752 | 0.041667 | 0.017857 | 0.02567 | 0.857329 | 0.853237 | 0.849888 | 0.836496 | 0.836496 | 0.836496 | 0 | 0.003535 | 0.362267 | 9,758 | 304 | 132 | 32.098684 | 0.860357 | 0.097356 | 0 | 0.85259 | 0 | 0 | 0.027411 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.103586 | false | 0.011952 | 0.015936 | 0.011952 | 0.227092 | 0.035857 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
2d9f81ee38a3d99745cc8cb0159333b13ca2f76e | 6,750 | py | Python | pirates/leveleditor/worldData/shipNavyMerchant3.py | Willy5s/Pirates-Online-Rewritten | 7434cf98d9b7c837d57c181e5dabd02ddf98acb7 | [
"BSD-3-Clause"
] | 81 | 2018-04-08T18:14:24.000Z | 2022-01-11T07:22:15.000Z | pirates/leveleditor/worldData/shipNavyMerchant3.py | Willy5s/Pirates-Online-Rewritten | 7434cf98d9b7c837d57c181e5dabd02ddf98acb7 | [
"BSD-3-Clause"
] | 4 | 2018-09-13T20:41:22.000Z | 2022-01-08T06:57:00.000Z | pirates/leveleditor/worldData/shipNavyMerchant3.py | Willy5s/Pirates-Online-Rewritten | 7434cf98d9b7c837d57c181e5dabd02ddf98acb7 | [
"BSD-3-Clause"
] | 26 | 2018-05-26T12:49:27.000Z | 2021-09-11T09:11:59.000Z | from pandac.PandaModules import Point3, VBase3, Vec4
objectStruct = {'Objects': {'1189040144.92gjeon': {'Type': 'Ship Part','Name': 'shipNavyMerchant3','Category': '13: War Galleon','File': '','Flagship': True,'Objects': {'1189040455.83gjeon': {'Type': 'Spawn Node','Aggro Radius': '12.0000','AnimSet': 'default','Hpr': Point3(0.0, 0.0, 0.0),'Min Population': '1','Patrol Radius': '12.0000','Pause Chance': '100','Pause Duration': '30','Pos': Point3(0.957, 13.753, 52.555),'Scale': VBase3(1.0, 1.0, 1.0),'Spawnables': 'Area','Start State': 'Patrol','Team': 'default','Visual': {'Color': (0, 0, 0.65, 1),'Model': 'models/misc/smiley'}},'1189040561.39gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-26.495, 45.455, 52.564),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040572.72gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-22.419, -3.694, 52.555),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040581.14gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(26.193, 46.064, 52.564),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040584.44gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(21.954, -5.448, 52.555),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040632.77gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-27.021, 84.795, 76.512),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040642.09gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-34.475, 115.512, 72.838),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040647.58gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(35.397, 116.217, 72.754),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040650.8gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': 100,'Pause Duration': 30,'Pos': Point3(27.643, 86.15, 76.35),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040698.55gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': 100,'Pause Duration': 30,'Pos': Point3(-21.446, -63.181, 94.129),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040703.08gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': 100,'Pause Duration': 30,'Pos': Point3(20.874, -63.84, 94.213),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040766.88gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-1.089, 51.312, 52.565),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189040838.08gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(0.676, 96.848, 75.059),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}},'1189041023.56gjeon': {'Type': 'Movement Node','Hpr': Point3(0.0, 0.0, 0.0),'Pause Chance': '100','Pause Duration': '30','Pos': Point3(-0.07, -10.163, 53.999),'Scale': VBase3(1.0, 1.0, 1.0),'Visual': {'Color': (0.65, 0, 0, 1),'Model': 'models/misc/smiley'}}},'Respawns': True,'Team': 'EvilNavy','Visual': {'Model': ['models/shipparts/merchantL3-geometry_High', 'models/shipparts/merchantL3-collisions', 'models/shipparts/merchantCabinAL3-collisions', 'models/shipparts/merchantCabinAL3-geometry_High']}}},'Node Links': [['1189040766.88gjeon', '1189040561.39gjeon', 'Bi-directional'], ['1189040766.88gjeon', '1189040581.14gjeon', 'Bi-directional'], ['1189040561.39gjeon', '1189040632.77gjeon', 'Bi-directional'], ['1189040642.09gjeon', '1189040632.77gjeon', 'Bi-directional'], ['1189040642.09gjeon', '1189040647.58gjeon', 'Bi-directional'], ['1189040650.8gjeon', '1189040647.58gjeon', 'Bi-directional'], ['1189040581.14gjeon', '1189040650.8gjeon', 'Bi-directional'], ['1189040581.14gjeon', '1189040647.58gjeon', 'Bi-directional'], ['1189040838.08gjeon', '1189040632.77gjeon', 'Bi-directional'], ['1189040650.8gjeon', '1189040838.08gjeon', 'Bi-directional'], ['1189040561.39gjeon', '1189040455.83gjeon', 'Bi-directional'], ['1189040581.14gjeon', '1189040455.83gjeon', 'Bi-directional'], ['1189040561.39gjeon', '1189040572.72gjeon', 'Bi-directional'], ['1189040581.14gjeon', '1189040584.44gjeon', 'Bi-directional'], ['1189040572.72gjeon', '1189040698.55gjeon', 'Bi-directional'], ['1189040703.08gjeon', '1189040584.44gjeon', 'Bi-directional'], ['1189040584.44gjeon', '1189041023.56gjeon', 'Bi-directional'], ['1189040572.72gjeon', '1189041023.56gjeon', 'Bi-directional']],'Layers': {},'ObjectIds': {'1189040144.92gjeon': '["Objects"]["1189040144.92gjeon"]','1189040455.83gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040455.83gjeon"]','1189040561.39gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040561.39gjeon"]','1189040572.72gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040572.72gjeon"]','1189040581.14gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040581.14gjeon"]','1189040584.44gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040584.44gjeon"]','1189040632.77gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040632.77gjeon"]','1189040642.09gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040642.09gjeon"]','1189040647.58gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040647.58gjeon"]','1189040650.8gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040650.8gjeon"]','1189040698.55gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040698.55gjeon"]','1189040703.08gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040703.08gjeon"]','1189040766.88gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040766.88gjeon"]','1189040838.08gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189040838.08gjeon"]','1189041023.56gjeon': '["Objects"]["1189040144.92gjeon"]["Objects"]["1189041023.56gjeon"]'}} | 3,375 | 6,697 | 0.657926 | 941 | 6,750 | 4.717322 | 0.167906 | 0.038297 | 0.038522 | 0.037846 | 0.434332 | 0.412255 | 0.391079 | 0.391079 | 0.3877 | 0.378914 | 0 | 0.272067 | 0.061778 | 6,750 | 2 | 6,697 | 3,375 | 0.428865 | 0 | 0 | 0 | 0 | 0 | 0.606429 | 0.16679 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.5 | 0 | 0.5 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 7 |
933663beb3b646af38f4e483e5c865f8634c1bfb | 16,451 | py | Python | tests/test__calculation.py | SteamPeKa/krippendorffs_alpha | c3d4f3eacaf418aeb22d30759594ca567dce9ecb | [
"MIT"
] | 1 | 2020-10-28T09:37:13.000Z | 2020-10-28T09:37:13.000Z | tests/test__calculation.py | SteamPeKa/krippendorffs_alpha | c3d4f3eacaf418aeb22d30759594ca567dce9ecb | [
"MIT"
] | null | null | null | tests/test__calculation.py | SteamPeKa/krippendorffs_alpha | c3d4f3eacaf418aeb22d30759594ca567dce9ecb | [
"MIT"
] | null | null | null | # coding=utf-8
# Creation date: 27 окт. 2020
# Creation time: 18:52
# Creator: SteamPeKa
import csv
import json
import os
import numpy
import pytest
import krippendorffs_alpha
import testing_utils
# Example E data matrix
OBSERVER_A_DATA = numpy.array([
[1, 0, 0, 0, 0, 1, 0, 1, 0, 0, 0, 0],
[0, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0],
[0, 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
]).T
OBSERVER_B_DATA = numpy.array([
[1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0],
[0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0],
[0, 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 1],
[0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0],
]).T
OBSERVER_C_DATA = numpy.array([
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0],
[0, 0, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0],
[0, 1, 1, 1, 0, 1, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0],
]).T
OBSERVER_D_DATA = numpy.array([
[1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 1, 0],
[0, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0],
[0, 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0],
]).T
DATA_MATRIX = numpy.zeros((4, 12, 5))
DATA_MATRIX[0, :, :] = OBSERVER_A_DATA
DATA_MATRIX[1, :, :] = OBSERVER_B_DATA
DATA_MATRIX[2, :, :] = OBSERVER_C_DATA
DATA_MATRIX[3, :, :] = OBSERVER_D_DATA
DATA_MATRIX.flags.writeable = False
OBSERVER_A_DATA.flags.writeable = False
OBSERVER_B_DATA.flags.writeable = False
OBSERVER_C_DATA.flags.writeable = False
OBSERVER_D_DATA.flags.writeable = False
class TestMakeCoincidencesMatrixFromDataMatrix(object):
def test_data_from_example_no_omit(self):
expected_value_by_unit_matrix = numpy.array([
[3, 0, 0, 0, 0, 1, 0, 3, 0, 0, 2, 0],
[0, 3, 0, 0, 4, 1, 0, 1, 4, 0, 0, 0],
[0, 1, 4, 4, 0, 1, 0, 0, 0, 0, 0, 1],
[0, 0, 0, 0, 0, 1, 4, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 3, 0, 0],
])
actual_value_by_unit_matrix = krippendorffs_alpha._calculation._make_value_by_unit_matrix_from_data_matrix(
DATA_MATRIX, omit_unpairable=False
)
testing_utils.assert_equal_tensors(expected_value_by_unit_matrix, actual_value_by_unit_matrix)
def test_data_from_example_omit_unpairable(self):
expected_value_by_unit_matrix = numpy.array([
[3, 0, 0, 0, 0, 1, 0, 3, 0, 0, 2],
[0, 3, 0, 0, 4, 1, 0, 1, 4, 0, 0],
[0, 1, 4, 4, 0, 1, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 1, 4, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 3, 0],
])
actual_value_by_unit_matrix = krippendorffs_alpha._calculation._make_value_by_unit_matrix_from_data_matrix(
DATA_MATRIX, omit_unpairable=True
)
testing_utils.assert_equal_tensors(expected_value_by_unit_matrix, actual_value_by_unit_matrix)
# noinspection PyPep8Naming
class Test_CalcAlpha(object):
def test_e_nominal(self):
metric_tensor = krippendorffs_alpha.metrics.NominalMetric().get_metric_tensor(list(range(1, 6)),
symmetric=False)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha(DATA_MATRIX, metric_tensor)
assert actual_alpha == pytest.approx(0.743, abs=0.001)
def test_e_interval(self):
metric_tensor = krippendorffs_alpha.metrics.IntervalMetric().get_metric_tensor(list(range(1, 6)),
symmetric=False)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha(DATA_MATRIX, metric_tensor)
assert actual_alpha == pytest.approx(0.849, abs=0.001)
class TestCalcAlphaFromListOrLists(object):
def test_e_nominal(self):
with open(os.path.join("tests", "example_E_data.tsv"), "r") as f:
input_table = csv.reader(f, delimiter="\t")
prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists(
input_table=input_table,
header=True,
row_legend=True,
upper_level="observer",
value_constructor=lambda s: int(s.strip()) if s.strip() != "NULL" else None
)
testing_utils.assert_equal_tensors(DATA_MATRIX, prepared_data.answers_tensor)
actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "nominal")
assert actual_alpha == pytest.approx(0.743, abs=0.001)
def test_e_interval(self):
with open(os.path.join("tests", "example_E_data.tsv"), "r") as f:
input_table = csv.reader(f, delimiter="\t")
prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists(
input_table=input_table,
header=True,
row_legend=True,
upper_level="observer",
value_constructor=lambda s: int(s.strip()) if s.strip() != "NULL" else None
)
testing_utils.assert_equal_tensors(DATA_MATRIX, prepared_data.answers_tensor)
actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "interval")
assert actual_alpha == pytest.approx(0.849, abs=0.001)
def test_wikipedia_nominal(self):
with open(os.path.join("tests", "example_wikipedia.csv"), "r") as f:
input_table = csv.reader(f, delimiter=",")
prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists(
input_table=input_table,
header=True,
row_legend=True,
upper_level="observer",
value_constructor=lambda s: int(s.strip()) if s.strip() != "*" else None
)
actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "nominal")
assert actual_alpha == pytest.approx(0.691, abs=0.001)
def test_wikipedia_interval(self):
with open(os.path.join("tests", "example_wikipedia.csv"), "r") as f:
input_table = csv.reader(f, delimiter=",")
prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists(
input_table=input_table,
header=True,
row_legend=True,
upper_level="observer",
value_constructor=lambda s: int(s.strip()) if s.strip() != "*" else None
)
actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "interval")
assert actual_alpha == pytest.approx(0.811, abs=0.001)
class TestCalcAlphaFromDictOfDicts(object):
def test_e_nominal(self):
with open(os.path.join("tests", "example_E_data.json"), "r") as f:
input_table = json.load(f)
prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts(
input_table=input_table,
upper_level="observer",
value_constructor=int
)
actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "nominal")
assert actual_alpha == pytest.approx(0.743, abs=0.001)
def test_e_interval(self):
with open(os.path.join("tests", "example_E_data.json"), "r") as f:
input_table = json.load(f)
prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts(
input_table=input_table,
upper_level="observer",
value_constructor=int
)
actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "interval")
assert actual_alpha == pytest.approx(0.849, abs=0.001)
def test_wikipedia_nominal(self):
with open(os.path.join("tests", "example_wikipedia.json"), "r") as f:
input_table = json.load(f)
prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts(
input_table=input_table,
upper_level="observer",
value_constructor=int
)
actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "nominal")
assert actual_alpha == pytest.approx(0.691, abs=0.001)
def test_wikipedia_interval(self):
with open(os.path.join("tests", "example_wikipedia.json"), "r") as f:
input_table = json.load(f)
prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts(
input_table=input_table,
upper_level="observer",
value_constructor=int
)
actual_alpha = krippendorffs_alpha._calculation.calc_alpha(prepared_data, "interval")
assert actual_alpha == pytest.approx(0.811, abs=0.001)
class Test_CalcAlphaByPrecomputesFromListOfLists(object):
def test_e_nominal(self):
metric_name = "nominal"
with open(os.path.join("tests", "example_E_data.tsv"), "r") as f:
input_table = csv.reader(f, delimiter="\t")
prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists(
input_table=input_table,
header=True,
row_legend=True,
upper_level="observer",
value_constructor=lambda s: int(s.strip()) if s.strip() != "NULL" else None
)
(assignment_matrix,
full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes(
prepared_data=prepared_data,
metric=metric_name
)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes(
assignment_matrix=assignment_matrix,
full_cross_disagreement_tensor=full_cross_disagreement_tensor
)
assert actual_alpha == pytest.approx(0.743, abs=0.001)
def test_e_interval(self):
metric_name = "interval"
with open(os.path.join("tests", "example_E_data.tsv"), "r") as f:
input_table = csv.reader(f, delimiter="\t")
prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists(
input_table=input_table,
header=True,
row_legend=True,
upper_level="observer",
value_constructor=lambda s: int(s.strip()) if s.strip() != "NULL" else None
)
(assignment_matrix,
full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes(
prepared_data=prepared_data,
metric=metric_name
)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes(
assignment_matrix=assignment_matrix,
full_cross_disagreement_tensor=full_cross_disagreement_tensor
)
assert actual_alpha == pytest.approx(0.849, abs=0.001)
def test_wikipedia_nominal(self):
metric_name = "nominal"
with open(os.path.join("tests", "example_wikipedia.csv"), "r") as f:
input_table = csv.reader(f, delimiter=",")
prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists(
input_table=input_table,
header=True,
row_legend=True,
upper_level="observer",
value_constructor=lambda s: int(s.strip()) if s.strip() != "*" else None
)
(assignment_matrix,
full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes(
prepared_data=prepared_data,
metric=metric_name
)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes(
assignment_matrix=assignment_matrix,
full_cross_disagreement_tensor=full_cross_disagreement_tensor
)
assert actual_alpha == pytest.approx(0.691, abs=0.001)
def test_wikipedia_interval(self):
metric_name = "interval"
with open(os.path.join("tests", "example_wikipedia.csv"), "r") as f:
input_table = csv.reader(f, delimiter=",")
prepared_data = krippendorffs_alpha.data_converters.from_list_of_lists(
input_table=input_table,
header=True,
row_legend=True,
upper_level="observer",
value_constructor=lambda s: int(s.strip()) if s.strip() != "*" else None
)
(assignment_matrix,
full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes(
prepared_data=prepared_data,
metric=metric_name
)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes(
assignment_matrix=assignment_matrix,
full_cross_disagreement_tensor=full_cross_disagreement_tensor
)
assert actual_alpha == pytest.approx(0.811, abs=0.001)
class Test_CalcAlphaByPrecomputesFromDictOfDicts(object):
def test_e_nominal(self):
metric_name = "nominal"
with open(os.path.join("tests", "example_E_data.json"), "r") as f:
input_table = json.load(f)
prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts(
input_table=input_table,
upper_level="observer",
value_constructor=int
)
(assignment_matrix,
full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes(
prepared_data=prepared_data,
metric=metric_name
)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes(
assignment_matrix=assignment_matrix,
full_cross_disagreement_tensor=full_cross_disagreement_tensor
)
assert actual_alpha == pytest.approx(0.743, abs=0.001)
def test_e_interval(self):
metric_name = "interval"
with open(os.path.join("tests", "example_E_data.json"), "r") as f:
input_table = json.load(f)
prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts(
input_table=input_table,
upper_level="observer",
value_constructor=int
)
(assignment_matrix,
full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes(
prepared_data=prepared_data,
metric=metric_name
)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes(
assignment_matrix=assignment_matrix,
full_cross_disagreement_tensor=full_cross_disagreement_tensor
)
assert actual_alpha == pytest.approx(0.849, abs=0.001)
def test_wikipedia_nominal(self):
metric_name = "nominal"
with open(os.path.join("tests", "example_wikipedia.json"), "r") as f:
input_table = json.load(f)
prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts(
input_table=input_table,
upper_level="observer",
value_constructor=int
)
(assignment_matrix,
full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes(
prepared_data=prepared_data,
metric=metric_name
)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes(
assignment_matrix=assignment_matrix,
full_cross_disagreement_tensor=full_cross_disagreement_tensor
)
assert actual_alpha == pytest.approx(0.691, abs=0.001)
def test_wikipedia_interval(self):
metric_name = "interval"
with open(os.path.join("tests", "example_wikipedia.json"), "r") as f:
input_table = json.load(f)
prepared_data = krippendorffs_alpha.data_converters.from_dict_of_dicts(
input_table=input_table,
upper_level="observer",
value_constructor=int
)
(assignment_matrix,
full_cross_disagreement_tensor) = krippendorffs_alpha._calculation._prepare_bootstrap_precomputes(
prepared_data=prepared_data,
metric=metric_name
)
actual_alpha = krippendorffs_alpha._calculation._calc_alpha_by_precomputes(
assignment_matrix=assignment_matrix,
full_cross_disagreement_tensor=full_cross_disagreement_tensor
)
assert actual_alpha == pytest.approx(0.811, abs=0.001)
| 41.860051 | 115 | 0.628412 | 2,094 | 16,451 | 4.618434 | 0.065425 | 0.046324 | 0.054906 | 0.060387 | 0.9301 | 0.915934 | 0.909317 | 0.906111 | 0.903733 | 0.903733 | 0 | 0.043212 | 0.2643 | 16,451 | 392 | 116 | 41.966837 | 0.755846 | 0.007781 | 0 | 0.740299 | 0 | 0 | 0.042658 | 0.010542 | 0 | 0 | 0 | 0 | 0.065672 | 1 | 0.059701 | false | 0 | 0.020896 | 0 | 0.098507 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
fa77928d2e1c2bfa9e2f7ccbfbd2167d176673e4 | 4,746 | py | Python | modules/MMM-toothbrush/test/BrushSessionTest.py | alexkahn139/MagicMirror | b988d85e833a79ebbdc453b562308056324aa983 | [
"MIT"
] | 13 | 2018-12-29T20:01:06.000Z | 2021-09-17T09:38:03.000Z | test/BrushSessionTest.py | alexkahn139/OralBlue_python | 5430ead8324872691eb5292af15d2e6fbeff9880 | [
"Apache-2.0"
] | null | null | null | test/BrushSessionTest.py | alexkahn139/OralBlue_python | 5430ead8324872691eb5292af15d2e6fbeff9880 | [
"Apache-2.0"
] | 8 | 2019-02-27T06:22:20.000Z | 2021-12-14T22:15:19.000Z | import unittest
from datetime import datetime, timedelta
from OralBlue import OralBAdvertise
from OralBlue.BrushMode import BrushMode
from OralBlue.BrushSession import BrushSession
from OralBlue.BrushState import BrushState
class BrushSessionTestCase(unittest.TestCase):
def test_anExceptionIsThrownWhenTheDataAreLessThan16Bytes(self):
with self.assertRaises(ValueError):
BrushSession(b"\x00")
def test_anExceptionIsThrownWhenTheDataAreMoreThan16Bytes(self):
with self.assertRaises(ValueError):
BrushSession(b"\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00")
def test_first4byteAreTheStartDate(self):
session = BrushSession(b"\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.startDate,datetime(year=2000,month=1,day=1))
session = BrushSession(b"\x0A\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.startDate, datetime(year=2000, month=1, day=1,second=10))
def test_byte5and6areTheDuration(self):
session = BrushSession(b"\x00\x01\x02\x03\x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.duration,timedelta(seconds=1))
session = BrushSession(b"\x00\x00\x00\x00\xb4\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.duration, timedelta(seconds=180))
session = BrushSession(b"\x00\x00\x00\x00\x00\x20\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.duration, timedelta(seconds=0x2000))
def test_byte8IsThePrefMode(self):
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x01\x00\x00\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.prefMode,BrushMode(0x01))
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x04\x00\x00\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.prefMode, BrushMode(0x04))
def test_byte9And10IsSecondsUnderPressure(self):
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x00\x00\x00\x00\x00")
self.assertEqual(session.timeUnderPressure,timedelta(seconds=0x0908))
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x01\x00\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.timeUnderPressure, timedelta(seconds=1))
def test_byte10IsNumberOfPressure(self):
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x00\x00\x00\x00\x00")
self.assertEqual(session.nPressure, 10)
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x00\x00\x00\x00\x00\x00")
self.assertEqual(session.nPressure,0)
def test_byte11IsBatteryCharge(self):
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x0B\x00\x00\x00\x00")
self.assertEqual(session.finalBatteryState, 11)
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x32\x00\x00\x00\x00")
self.assertEqual(session.finalBatteryState,50)
def test_last4BytesAreTheLastCharge(self):
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x0B\x00\x00\x00\x00")
self.assertEqual(session.lastCharge,datetime(year=2000,month=1,day=1))
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x3B\x0A\x00\x00\x00")
self.assertEqual(session.lastCharge,datetime(year=2000, month=1, day=1,second=10))
class BrushSessionV2Or3TestCase(unittest.TestCase):
def test_bytes12AsNSectionAndTargetTime(self):
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x0B\x00\x00\x00\x00",protocolVersion=3)
self.assertEqual(session.lastCharge,None)
self.assertEqual(session.numberOfSector,0)
self.assertEqual(session.sessionTargetTime, 0)
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x3B\x78\x80\x00\x00",protocolVersion=3)
self.assertEqual(session.lastCharge, None)
self.assertEqual(session.numberOfSector, 4)
self.assertEqual(session.sessionTargetTime, 120)
def test_bytes14AsSessionIdAndUserId(self):
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x0B\x0C\x0D\x00\x00",protocolVersion=3)
self.assertEqual(session.lastCharge,None)
self.assertEqual(session.sessionId,0)
self.assertEqual(session.userId, 0)
session = BrushSession(b"\x00\x01\x02\x03\x04\x05\x06\x07\x08\x09\x0A\x3B\x78\x80\x03\x20",protocolVersion=3)
self.assertEqual(session.lastCharge, None)
self.assertEqual(session.sessionId, 3)
self.assertEqual(session.userId, 1)
if __name__ == '__main__':
unittest.main()
| 53.325843 | 117 | 0.722714 | 657 | 4,746 | 5.191781 | 0.135464 | 0.228672 | 0.287599 | 0.316623 | 0.734682 | 0.733216 | 0.732337 | 0.722955 | 0.654647 | 0.631193 | 0 | 0.173723 | 0.134008 | 4,746 | 88 | 118 | 53.931818 | 0.656204 | 0 | 0 | 0.140845 | 0 | 0.28169 | 0.273072 | 0.270544 | 0 | 0 | 0.004214 | 0 | 0.408451 | 1 | 0.15493 | false | 0 | 0.084507 | 0 | 0.267606 | 0 | 0 | 0 | 0 | null | 1 | 1 | 1 | 0 | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | null | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 9 |
fac00226a22a7fdcd7612abc634901a3d3a2425a | 162 | py | Python | peerplaysapi/wallet.py | farisshajahan/python-peerplays | 260e43d39307554f5c35574fe7a6f5451a51697f | [
"MIT"
] | 10 | 2019-03-14T03:09:42.000Z | 2021-03-17T10:20:05.000Z | peerplaysapi/wallet.py | farisshajahan/python-peerplays | 260e43d39307554f5c35574fe7a6f5451a51697f | [
"MIT"
] | 8 | 2019-04-02T17:07:20.000Z | 2020-04-30T08:24:01.000Z | peerplaysapi/wallet.py | farisshajahan/python-peerplays | 260e43d39307554f5c35574fe7a6f5451a51697f | [
"MIT"
] | 4 | 2019-04-02T17:00:22.000Z | 2021-08-09T11:28:20.000Z | from grapheneapi.grapheneapi import GrapheneAPI
class PeerPlaysWalletRPC(GrapheneAPI):
""" This class inherits everything from GrapheneAPI
"""
pass
| 20.25 | 55 | 0.759259 | 15 | 162 | 8.2 | 0.6 | 0.243902 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.179012 | 162 | 7 | 56 | 23.142857 | 0.924812 | 0.290123 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0.333333 | 0.333333 | 0 | 0.666667 | 0 | 1 | 0 | 0 | null | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 7 |
fad2774df466156d62a3a38ab65ff7e2b6897589 | 146 | py | Python | angrcli/full/__init__.py | cokesme/angr-cli-mirror | 37da5bf731549dab2523215a9212f372e3302168 | [
"MIT"
] | 38 | 2018-10-18T17:03:52.000Z | 2022-02-06T13:19:19.000Z | angrcli/full/__init__.py | cokesme/angr-cli-mirror | 37da5bf731549dab2523215a9212f372e3302168 | [
"MIT"
] | 3 | 2019-06-29T10:16:20.000Z | 2021-06-10T15:29:27.000Z | angrcli/full/__init__.py | cokesme/angr-cli-mirror | 37da5bf731549dab2523215a9212f372e3302168 | [
"MIT"
] | 4 | 2018-10-08T12:54:50.000Z | 2020-03-27T09:29:27.000Z | from angrcli.plugins.watches import Watches
from angrcli.plugins.ContextView import ContextView
from angrcli.plugins.explore import ExplorePlugin
| 36.5 | 51 | 0.876712 | 18 | 146 | 7.111111 | 0.444444 | 0.257813 | 0.421875 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.082192 | 146 | 3 | 52 | 48.666667 | 0.955224 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 7 |
fada392eade7c3cd7bc3b09d1b80c50aa2b5674b | 44,679 | py | Python | tests/integration_tests/test_resources/new_test_ads.py | JobtechSwe/elastic-importers | ae370984f79295a784350f98e695977a1f73647a | [
"Apache-2.0"
] | 2 | 2020-03-02T18:29:08.000Z | 2021-06-09T00:48:24.000Z | tests/integration_tests/test_resources/new_test_ads.py | JobtechSwe/elastic-importers | ae370984f79295a784350f98e695977a1f73647a | [
"Apache-2.0"
] | 12 | 2019-01-31T09:54:23.000Z | 2021-10-04T11:25:44.000Z | tests/integration_tests/test_resources/new_test_ads.py | JobtechSwe/elastic-importers | ae370984f79295a784350f98e695977a1f73647a | [
"Apache-2.0"
] | 1 | 2018-11-09T14:44:55.000Z | 2018-11-09T14:44:55.000Z | test_ads = [
{
"annonsId": 23448227,
"annonsrubrik": "Driven fältsäljare Östergötland/Västergötland",
"annonstext": "Köksbörsen är inne i ett expansivt skede. Vi har egen import ifrån Europa och Asien samt USA. Vi har egna agenturprodukter. Vi levererar och projekterar kompletta restaurangkök över hela Sverige. Vi har en säljare i Göteborg, en säljare i Umeå/Skellefteå samt ett antal säljare i Stockholm. Vi expanderar nu säljteamet i Stockholm med ytterligare en fältsäljare för Stockholm/Mälardalen, då vi har sett ett utökat behov av ytterligare en säljare, vi ett härligt säljteam!\nVi ligger centralt i Stockholm vid Gärdet/Östermalm. VI har egen utställning av nya/begagnade restaurangmaskiner, samt även en utställning för möbler/glas/porslin/bestick. För att lyckas med denna tjänst, så krävs flera års erfarenhet inom försäljning till dagligvaruhandeln/restaurangbranschen, det krävs att du har ett brett kontaktnät inom branschen och många kontakter att börja jobba med direkt. Det är ett stort plus om du tidigare har jobbat som säljare av restaurangmaskiner/inredning till restauranger/dagligvaruhandeln. I den här rollen kommer fokus att ligga på nykundsbearbetning, men du kommer även att utveckla befintliga kunder. Du ansvarar för din egen budget, kundstock och du bokar dina egna möten i ditt arbete.\nDet är meriterande om du har goda försäljningsresultat sedan tidigare och det är ett krav att du har körkort. För att lyckas i rollen bör du tycka om att arbeta med uppsökande försäljning och vara duktig på att skapa ett brett kontaktnät inom branschen. Du ska självständigt kunna planera ditt arbete och vara en driven person som är hungrig på nya utmaningar.\nDu är en självgående person som kan ta egna initiativ till nya affärer. Som person är du driven, stresstålig och ansvarsfull. Du har skinn på näsan och kan hantera många saker samtidigt. Det är viktigt att du är målinriktad och resultatfokuserad i ditt arbete och vill få snabba avslut i dina affärer. Vi ser att du är en social person som tycker om att ha mycket kundkontakt samt har ett sinne för att göra affärer. Vi ser det som meriterande om du har CAD vana. En fördel är om du är utbildad kock, eller har jobbat på restaurang. Vi söker en driven fältsäljare som kommer att bearbeta nya samt befintliga kunder i Östergötland/Västergötland. Du ansvarar för din egen budget, kundstock och du bokar dina egna möten. I rollen som fältsäljare planerar du dina egna dagar och det är därför viktigt att du är ansvarsfull och målinriktad i ditt arbete.\nFör att lyckas med denna tjänst, så krävs flera års erfarenhet inom försäljning till dagligvaruhandeln/restaurangbranschen, det krävs att du har ett brett kontaktnät inom branschen och många kontakter att börja jobba med direkt. Du kommer även jobba med ringlistor för att bearbeta kedjekunder samt byggprojekt. Dom kunder som du ska bearbeta är restaurangkedjor, hotel, byggfirmor, kökskonsulter/arkitekter, livsmedelsbutiker, bensinstationer mm.",
"annonstextFormaterad": "Köksbörsen är inne i ett expansivt skede. Vi har egen import ifrån Europa och Asien samt USA. Vi har egna agenturprodukter. Vi levererar och projekterar kompletta restaurangkök över hela Sverige. Vi har en säljare i Göteborg, en säljare i Umeå/Skellefteå samt ett antal säljare i Stockholm. Vi expanderar nu säljteamet i Stockholm med ytterligare en fältsäljare för Stockholm/Mälardalen, då vi har sett ett utökat behov av ytterligare en säljare, vi ett härligt säljteam!<br>Vi ligger centralt i Stockholm vid Gärdet/Östermalm. VI har egen utställning av nya/begagnade restaurangmaskiner, samt även en utställning för möbler/glas/porslin/bestick. För att lyckas med denna tjänst, så krävs flera års erfarenhet inom försäljning till dagligvaruhandeln/restaurangbranschen, det krävs att du har ett brett kontaktnät inom branschen och många kontakter att börja jobba med direkt. Det är ett stort plus om du tidigare har jobbat som säljare av restaurangmaskiner/inredning till restauranger/dagligvaruhandeln. I den här rollen kommer fokus att ligga på nykundsbearbetning, men du kommer även att utveckla befintliga kunder. Du ansvarar för din egen budget, kundstock och du bokar dina egna möten i ditt arbete.<br>Det är meriterande om du har goda försäljningsresultat sedan tidigare och det är ett krav att du har körkort. För att lyckas i rollen bör du tycka om att arbeta med uppsökande försäljning och vara duktig på att skapa ett brett kontaktnät inom branschen. Du ska självständigt kunna planera ditt arbete och vara en driven person som är hungrig på nya utmaningar.<br>Du är en självgående person som kan ta egna initiativ till nya affärer. Som person är du driven, stresstålig och ansvarsfull. Du har skinn på näsan och kan hantera många saker samtidigt. Det är viktigt att du är målinriktad och resultatfokuserad i ditt arbete och vill få snabba avslut i dina affärer. Vi ser att du är en social person som tycker om att ha mycket kundkontakt samt har ett sinne för att göra affärer. Vi ser det som meriterande om du har CAD vana. En fördel är om du är utbildad kock, eller har jobbat på restaurang. Vi söker en driven fältsäljare som kommer att bearbeta nya samt befintliga kunder i Östergötland/Västergötland. Du ansvarar för din egen budget, kundstock och du bokar dina egna möten. I rollen som fältsäljare planerar du dina egna dagar och det är därför viktigt att du är ansvarsfull och målinriktad i ditt arbete.<br>För att lyckas med denna tjänst, så krävs flera års erfarenhet inom försäljning till dagligvaruhandeln/restaurangbranschen, det krävs att du har ett brett kontaktnät inom branschen och många kontakter att börja jobba med direkt. Du kommer även jobba med ringlistor för att bearbeta kedjekunder samt byggprojekt. Dom kunder som du ska bearbeta är restaurangkedjor, hotel, byggfirmor, kökskonsulter/arkitekter, livsmedelsbutiker, bensinstationer mm.",
"ansokningssattEpost": " EMAIL ",
"ansokningssattViaAF": False,
"anstallningTyp": {
"namn": "Vanlig anställning",
"varde": "PFZr_Syz_cUq"
},
"antalPlatser": 1,
"arbetsgivareId": "20513633",
"arbetsgivareNamn": "Köksbörsen AB",
"arbetsplatsId": "83191774",
"arbetsplatsNamn": "KÖKSBÖRSEN AB",
"arbetsplatsadress": {
"gatuadress": "Linköping/Jönköping",
"kommun": {
"namn": "Linköping",
"varde": "bm2x_1mr_Qhx"
},
"koordinatPrecision": "POSTNUMMER",
"lan": {
"namn": "Östergötlands län",
"varde": "oLT3_Q9p_3nn"
},
"land": {
"namn": "Sverige",
"varde": "i46j_HmG_v64"
},
"latitud": "58.4044001929687",
"longitud": "15.6003742645864",
"postnr": "58002",
"postort": "LINKÖPING"
},
"arbetstidTyp": {
"namn": "Heltid",
"varde": "6YE1_gAC_R2G"
},
"avpublicerad": False,
"ejKravSvenskaEngelska": False,
"informationAnsokningssatt": None,
"ingenErfarenhetKravs": False,
"inkluderande": False,
"kallaTyp": "VIA_ANNONSERA",
"kontaktpersoner": [
{
"befattning": None,
"beskrivning": "blabla",
"efternamn": "Testsson",
"epost": "test@jobtechdev.se",
"fackligRepresentant": False,
"fornamn": "Testy",
"telefonnummer": "+01011122233"
}
],
"lonTyp": {
"namn": "Fast och rörlig lön",
"varde": "asrX_9Df_ukn"
},
"organisationsnummer": "5567425631",
"publiceringsdatum": "2021-03-19 12:41:52",
"referens": "Utesäljare",
"sistaAnsokningsdatum": "2021-04-18 23:59:59",
"sistaPubliceringsdatum": "2022-06-14 07:38:46",
"telefonnummer": None,
"tillgangTillEgenBil": False,
"uppdateradTid": 1616154112141,
"utbildningsinriktning": {
"namn": "Företagsekonomi, handel och administration",
"varde": "CU11_PU7_Yj1",
"vikt": 5
},
"utbildningsniva": {
"namn": "Gymnasial utbildning",
"varde": "DeBt_ahh_bkx",
"vikt": 5
},
"varaktighetTyp": {
"namn": "Tillsvidare",
"varde": "a7uU_j21_mkL"
},
"version": "1.0",
"yrkeserfarenheter": [
{
"erfarenhetsniva": {
"namn": "1-2 års erfarenhet",
"varde": "LLnd_5GJ_4ju"
},
"namn": "Utesäljare",
"varde": "p17k_znk_osi",
"vikt": 10
}
],
"yrkesroll": {
"namn": "Utesäljare",
"varde": "p17k_znk_osi"
}
},
{'annonsId': 23699999, 'annonsrubrik': 'Greenkeeper -- Banpersonal 2021 Ullna Golf',
'annonstext': 'Ullna Golf Club är en klassisk golfanläggning som funnits sedan 1981, vi har som mål att presentera en förstklassig golfbana och erbjuda en service med hög nivå.\n\nDu som söker dig till oss är positiv, flexibel, ordningsam och intresserad av att jobba i grupp. Du kommer jobba med skötsel av golfbanan. Det är tidiga morgnar, runt 06.00, i gengäld slutar vi tidigt på eftermiddagen, ca 15.00, du kommer även jobba några helger.\n\nDina arbetsuppgifter kommer till största del bestå av att klippa tees (utslagsplatser), greener (mest finklippta ytan), klippa ruffar samt kratta bunkrar.\n\nVi söker ett flertal banarbetare med varierande start av tjänsterna. De första tjänsterna börjar i FIRSTNAME/LASTNAME och löper i ca 4-6 månader beroende på tjänst. Det är ett krav att du kan ta dig hit själv, vi har mycket dåliga kollektiva kommunikationer. Du måste även kunna göra dig förstådd och förstå engelska. Det är meriterande om du spelar golf och/eller har jobbat på en golfbana, har B-kort eller traktorbehörighet. Det är bra om du har jobbat med röj såg, kört traktor.\n\nVi ser fram emot att läsa din ansökan och i slutändan jobba tillsammans.\n\nRekryteringen sker löpande.\n\nFör frågor och ansökningar mailas till \n\n Banchef FIRSTNAME LASTNAME\n\n EMAIL \n\nSkriv "Banpersonal2021" i ämnesraden\n\nDu kan besöka oss på Ullnagolf.se',
'annonstextFormaterad': '<p>Ullna Golf Club är en klassisk golfanläggning som funnits sedan 1981, vi har som mål att presentera en förstklassig golfbana och erbjuda en service med hög nivå.</p><p>Du som söker dig till oss är positiv, flexibel, ordningsam och intresserad av att jobba i grupp. Du kommer jobba med skötsel av golfbanan. Det är tidiga morgnar, runt 06.00, i gengäld slutar vi tidigt på eftermiddagen, ca 15.00, du kommer även jobba några helger.</p><p>Dina arbetsuppgifter kommer till största del bestå av att klippa tees (utslagsplatser), greener (mest finklippta ytan), klippa ruffar samt kratta bunkrar.</p><p>Vi söker ett flertal banarbetare med varierande start av tjänsterna. De första tjänsterna börjar i FIRSTNAME/LASTNAME och löper i ca 4-6 månader beroende på tjänst. Det är ett krav att du kan ta dig hit själv, vi har mycket dåliga kollektiva kommunikationer. Du måste även kunna göra dig förstådd och förstå engelska. Det är meriterande om du spelar golf och/eller har jobbat på en golfbana, har B-kort eller traktorbehörighet. Det är bra om du har jobbat med röj såg, kört traktor.</p><p>Vi ser fram emot att läsa din ansökan och i slutändan jobba tillsammans.</p><p>Rekryteringen sker löpande.</p><p><br></p><p>För frågor och ansökningar mailas till </p><p> Banchef FIRSTNAME LASTNAME</p><p> EMAIL </p><p>Skriv "Banpersonal2021" i ämnesraden</p><p>Du kan besöka oss på Ullnagolf.se</p><p><br></p><p><br></p>',
'ansokningssattEpost': ' EMAIL ', 'ansokningssattViaAF': False,
'anstallningTyp': {'namn': 'Sommarjobb / feriejobb', 'varde': 'Jh8f_q9J_pbJ'}, 'antalPlatser': 4,
'arbetsgivareId': '21057934', 'arbetsgivareNamn': 'Ullna Golf AB', 'arbetsplatsId': '85992110',
'arbetsplatsNamn': 'Ullna Golf AB',
'arbetsplatsadress': {'gatuadress': 'Roslagsvägen 36', 'kommun': {'namn': 'Österåker', 'varde': '8gKt_ZsV_PGj'},
'koordinatPrecision': 'GATUADRESS',
'lan': {'namn': 'Stockholms län', 'varde': 'CifL_Rzy_Mku'},
'land': {'namn': 'Sverige', 'varde': 'i46j_HmG_v64'}, 'latitud': '59.49548570116787',
'longitud': '18.152517169298417', 'postnr': '18494', 'postort': 'Åkersberga'},
'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'ejKravSvenskaEngelska': False,
'informationAnsokningssatt': None, 'ingenErfarenhetKravs': False, 'inkluderande': False,
'kallaTyp': 'VIA_ANNONSERA', 'kontaktpersoner': [
{'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se',
'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}],
'korkort': [{'namn': 'B', 'varde': 'VTK8_WRx_GcM'}],
'lonTyp': {'namn': 'Fast månads- vecko- eller timlön', 'varde': 'oG8G_9cW_nRf'},
'organisationsnummer': '5560428095', 'publiceringsdatum': '2021-03-18 00:00:00', 'referens': 'Banpersonal 2021',
'sistaAnsokningsdatum': '2021-04-16 23:59:59', 'sistaPubliceringsdatum': '2022-06-14 14:33:12',
'telefonnummer': None, 'tillgangTillEgenBil': True, 'uppdateradTid': 1615971812090,
'varaktighetTyp': {'namn': '3 - 6 månader', 'varde': 'Xj7x_7yZ_jEn'}, 'version': '1.0',
'webbadress': 'ullnagolf.se', 'yrkesroll': {'namn': 'Golfbanearbetare', 'varde': 'tArV_EVU_cFQ'}},
{'annonsId': 24277938, 'annonsrubrik': 'Sjuksköterska till vårdplatsenheten',
'annonstext': 'Kullbergska sjukhuset Katrineholm \r\n\r\nVälkommen att skapa Sveriges friskaste län tillsammans med oss!\n\nSjuksköterska till vårdplatsenheten, se hit!\n\nOm oss\nVårdplatsenheten har två vårdavdelningar varav den ena med inriktning hjärt -och medicin och den andra stroke- och medicin. Vårdplatsenheten består av ett härligt gäng medarbetare som nu behöver utöka sjuksköterskegruppen. \nAvdelningarna vårdar patienter med olika sjukdomstillstånd, vi söker dig som är intresserad av hjärtsjukvård och strokevård/ortopedrehab.\nDu ska vara nyfiken, flexibel och kunna anpassa dig till ett varierande arbete och tempo.\n\nVi jobbar i team bestående av sjuksköterskor, undersköterskor, utskrivningssamordnare och medicinska sekreterare. Läkare och paramedicinsk kompetens finns på avdelningen under vissa tider som stöd i arbetet. \n\nTillsammans arbetar vi målinriktat för att skapa en trygg och säker resa genom vården för våra kunder.\n\nDin kompetens\nLegitimerad sjuksköterska. Erfarenhet av slutenvård är meriterande men inget krav. Vi söker en flexibel person som uppskattar en varierad arbetsdag med möjlighet att lära sig mycket i arbetet. Du ska ha lätt att samarbeta med olika professioner, och vara öppen för förändringar. Vi lägger stor vikt på att vara lyhörd och göra insatser där behovet är som störst. Personliga egenskaper lägger vi stor vikt vid.\n\nAnställningsform\nTillsvidareanställning / Vikariat på heltid eller enligt överenskommelse. Tillträde enligt överenskommelse.\nArbetstid dag/kväll/helg eller natt. Planeringsschema används på avdelningen med möjlighet att kunna påverka sina pass till viss del.\n\nInformation om tjänsten lämnas av\nVårdenhetschef FIRSTNAME LASTNAME (stroke- och medicin), TELEPHONENO .\nVårdenhetschef Ann-FIRSTNAME LASTNAME (hjärtmedicin), TELEPHONENO .\nFacklig företrädare Arleide Pereira FIRSTNAME LASTNAME, TELEPHONENO och FIRSTNAME LASTNAME, TELEPHONENO .\nÖvriga fackliga företrädare nås via kontaktcenter, TELEPHONENO .\n\nKom och jobba hos oss på vårdplatsenheten!\n\nVälkommen med din ansökan, inklusive CV, senast 2021-03-29.\nIntervjuer kan komma att ske löpande.\n\nSe våra förmåner (http://regionsormland.se/jobb-och-utbildning/vara-formaner/)\n\nFölj oss gärna på Facebook (http://facebook.com/regionsormlandJobbahososs/)\n\r\n\nVi har i vissa fall skyldighet att kontrollera om en person förekommer i misstanke- och eller belastningsregistret. Det kan ske på två sätt, endera begär regionen ut uppgiften själv eller också uppmanas du att begära ut utdrag för att kunna visa upp innan anställning. Vi begär i undantagsfall att du visar upp registerutdrag även vid tillsättning av andra tjänster än de som avses ovan. Blir du aktuell för anställning kommer du att informeras om vad som gäller för den tjänst du sökt.\n\nVi gör aktiva val vid exponering och rekryteringsstöd och undanber oss därför direktkontakt av bemannings- och rekryteringsföretag.',
'annonstextFormaterad': 'Kullbergska sjukhuset Katrineholm \r\n\r\nVälkommen att skapa Sveriges friskaste län tillsammans med oss!\n\nSjuksköterska till vårdplatsenheten, se hit!\n\nOm oss\nVårdplatsenheten har två vårdavdelningar varav den ena med inriktning hjärt -och medicin och den andra stroke- och medicin. Vårdplatsenheten består av ett härligt gäng medarbetare som nu behöver utöka sjuksköterskegruppen. \nAvdelningarna vårdar patienter med olika sjukdomstillstånd, vi söker dig som är intresserad av hjärtsjukvård och strokevård/ortopedrehab.\nDu ska vara nyfiken, flexibel och kunna anpassa dig till ett varierande arbete och tempo.\n\nVi jobbar i team bestående av sjuksköterskor, undersköterskor, utskrivningssamordnare och medicinska sekreterare. Läkare och paramedicinsk kompetens finns på avdelningen under vissa tider som stöd i arbetet. \n\nTillsammans arbetar vi målinriktat för att skapa en trygg och säker resa genom vården för våra kunder.\n\nDin kompetens\nLegitimerad sjuksköterska. Erfarenhet av slutenvård är meriterande men inget krav. Vi söker en flexibel person som uppskattar en varierad arbetsdag med möjlighet att lära sig mycket i arbetet. Du ska ha lätt att samarbeta med olika professioner, och vara öppen för förändringar. Vi lägger stor vikt på att vara lyhörd och göra insatser där behovet är som störst. Personliga egenskaper lägger vi stor vikt vid.\n\nAnställningsform\nTillsvidareanställning / Vikariat på heltid eller enligt överenskommelse. Tillträde enligt överenskommelse.\nArbetstid dag/kväll/helg eller natt. Planeringsschema används på avdelningen med möjlighet att kunna påverka sina pass till viss del.\n\nInformation om tjänsten lämnas av\nVårdenhetschef FIRSTNAME LASTNAME (stroke- och medicin), TELEPHONENO .\nVårdenhetschef Ann-FIRSTNAME LASTNAME (hjärtmedicin), TELEPHONENO .\nFacklig företrädare Arleide Pereira FIRSTNAME LASTNAME, TELEPHONENO och FIRSTNAME LASTNAME, TELEPHONENO .\nÖvriga fackliga företrädare nås via kontaktcenter, TELEPHONENO .\n\n\nKom och jobba hos oss på vårdplatsenheten!\n\n\nVälkommen med din ansökan, inklusive CV, senast 2021-03-29.\nIntervjuer kan komma att ske löpande.\n\nSe våra förmåner (http://regionsormland.se/jobb-och-utbildning/vara-formaner/)\n\nFölj oss gärna på Facebook (http://facebook.com/regionsormlandJobbahososs/)\n\r\n\nVi har i vissa fall skyldighet att kontrollera om en person förekommer i misstanke- och eller belastningsregistret. Det kan ske på två sätt, endera begär regionen ut uppgiften själv eller också uppmanas du att begära ut utdrag för att kunna visa upp innan anställning. Vi begär i undantagsfall att du visar upp registerutdrag även vid tillsättning av andra tjänster än de som avses ovan. Blir du aktuell för anställning kommer du att informeras om vad som gäller för den tjänst du sökt.\n\nVi gör aktiva val vid exponering och rekryteringsstöd och undanber oss därför direktkontakt av bemannings- och rekryteringsföretag.',
'ansokningsadress': {'gatuadress': 'Repslagaregatan 19', 'land': None, 'mottagare': None, 'postnr': None,
'postort': None}, 'ansokningssattEpost': None, 'ansokningssattViaAF': False,
'ansokningssattWebbadress': 'https://sormland.powerinit.com/Modules/Recruitments/Public/?JobPositionId=18351&RefNo=RLSV-20-564&Source=[JobPositionSource]',
'anstallningTyp': {'namn': 'Vanlig anställning', 'varde': 'PFZr_Syz_cUq'}, 'antalPlatser': 1,
'arbetsgivareId': '10820079', 'arbetsgivareNamn': 'REGION SÖRMLAND', 'arbetsplatsId': '0',
'arbetsplatsNamn': 'Region Sörmland',
'arbetsplatsadress': {'gatuadress': None, 'kommun': {'namn': 'Katrineholm', 'varde': 'snx9_qVD_Dr1'},
'koordinatPrecision': 'KOMMUN',
'lan': {'namn': 'Södermanlands län', 'varde': 's93u_BEb_sx2'},
'land': {'namn': 'Sverige', 'varde': 'i46j_HmG_v64'}, 'latitud': '58.995552',
'longitud': '16.205475', 'postnr': None, 'postort': None},
'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False,
'besoksadress': {'gatuadress': 'Repslagaregatan 19', 'land': None, 'postnr': None, 'postort': None},
'ejKravSvenskaEngelska': False, 'externtAnnonsId': '46-232100-0032-18351', 'informationAnsokningssatt': None,
'ingenErfarenhetKravs': False, 'inkluderande': False, 'kallaTyp': 'VIA_PLATSBANKEN_DXA', 'kontaktpersoner': [
{'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se',
'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}],
'lonTyp': {'namn': 'Fast månads- vecko- eller timlön', 'varde': 'oG8G_9cW_nRf'}, 'lonebeskrivning': '-',
'organisationsnummer': '2321000032',
'postadress': {'gatuadress': 'Repslagaregatan 19', 'land': 'SE', 'postnr': '61188', 'postort': 'Nyköping'},
'publiceringsdatum': '2020-09-30 14:35:20', 'referens': 'RLSV-20-564',
'sistaAnsokningsdatum': '2021-03-29 23:59:59', 'sistaPubliceringsdatum': '2022-06-14 14:33:59',
'telefonnummer': None, 'tillgangTillEgenBil': False, 'uppdateradTid': 1613993432534,
'varaktighetTyp': {'namn': 'Tillsvidare', 'varde': 'a7uU_j21_mkL'}, 'version': '4.0',
'villkorsbeskrivning': 'Tillsvidare, 100, Tillträde enligt överenskommelse\r\n\n-', 'yrkeserfarenheter': [
{'erfarenhetsniva': {'namn': 'Mindre än 1 års erfarenhet', 'varde': 'yrAe_Fzi_E6u'},
'namn': 'Sjuksköterska, grundutbildad', 'varde': 'bXNH_MNX_dUR', 'vikt': 10}],
'yrkesroll': {'namn': 'Sjuksköterska, grundutbildad', 'varde': 'bXNH_MNX_dUR'}}]
ads_not_sweden = [
{'annonsId': 24458381, 'annonsrubrik': 'Servicetekniker i Oslo sökes',
'annonstext': 'Er du en handy og serviceinnstilt person som ønsker en spennende stilling i en fremoverlent og voksende bedrift?\n\nSom servicetekniker vil du ha ansvar for service og hjemmebesøk for Part Construction AB. Du vil behandle innkommende forespørsler, utføre service oppdrag på byggeprosjekter og reklamasjonsarbeid. Arbeidet foregår hovedsakelig på Østlandet, men vi har også leveranser til andre deler av landet. Noe reising vil derfor måtte påberegnes.\n\nOm deg:\n\n\tEn allsidig praktisk fagarbeider. Du er kanskje flislegger, rørlegger, elektriker eller har en annen relevant utdanning\n\tSnakker og skriver godt Norsk\n\tFørerkort klasse B/BE\n\tGod evne til kundebehandling\n\tEr selvstendig og strukturert\n\tGrunnleggende data kunnskaper.\n\tKjennskap til NS 3420, NS 8409 og Bustadoppføringslova.\n\nPart Construction AB kan tilby:\n\n\tEt dynamisk og spennende arbeidsmiljø\n\tUtfordrende og variert arbeidsoppgaver\n\tKonkurransedyktig betingelser\n\tServicebil, mobil telefon\n\tGode forsikringer og pensjons ordninger\n\nOm arbeidsgiveren \n\nPart Construction AB, en av Europas ledende produsenter av prefabrikkerte baderoms moduler. Part Construction AB tilbyr entreprenørfirmaer og eiendomsselskaper prefabrikkerte baderom til blant annet, hoteller, sykehjem og boliger. Bademodulene leveres komplett med fliser, gulv, møbler, inventar, helt etter kundens ønsker. rørleggerarbeid og elektriske installasjoner er også ferdig montert fra fabrikk, Part er et familieeid selskap fra Kalix, med ca. 300 ansatte og med en omsetning på ca. 600 millioner. Hovedkvarter og produksjon er i Kalix / Överkalix, Norrbotten. Part Construction AB er en del av konserngruppen PartGroup sammen med søsterselskapene PreBad AB, Altor Industri, Isolamin AB, Space Interior AB, og PCS Modulsystem.\n\nSektor\n\nPrivat\n\nSted\n\nVestbygata 55 2003 Lillestrøm\n\nBransje\n\nBygg og anlegg,\n\nStillingsfunksjon\n\nHåndverker / Andre montører,\n\nHåndverker,\n\nHåndverker / Flislegger, rørlegger, elektriker\n\n \n\nSøknad\n\nTiltredelse\n\nEtter avtale\n\nVarighet\n\nFast, heltid (6 måneders prøvetid)\n\nBosted\n\nOslo området\n\nSøknadsfrist\n\nSnarest, og innen 10.03.2021\n\nSpørsmål om stillingen rettes til FIRSTNAME LASTNAME Yttervik, Servicekoordinator Norge. Tel: TELEPHONENO , e-post: EMAIL \n\nSend søknad merket "Servicetekniker Norge" på e-post til: EMAIL \n\nDenne rekrutteringen skjer helt i Part Constructions regi, og vi ønsker kun direkte kontakt med personlig søkere. \n\nVi gleder oss til å høre fra deg!',
'annonstextFormaterad': '<p><strong>Er du en handy og serviceinnstilt person som ønsker en spennende stilling i en fremoverlent og voksende bedrift?</strong></p><p>Som servicetekniker vil du ha ansvar for service og hjemmebesøk for Part Construction AB. Du vil behandle innkommende forespørsler, utføre service oppdrag på byggeprosjekter og reklamasjonsarbeid. Arbeidet foregår hovedsakelig på Østlandet, men vi har også leveranser til andre deler av landet. Noe reising vil derfor måtte påberegnes.</p><p><br></p><p><strong>Om deg:</strong></p><ul><li>En allsidig praktisk fagarbeider. Du er kanskje flislegger, rørlegger, elektriker eller har en annen relevant utdanning</li><li>Snakker og skriver godt Norsk</li><li>Førerkort klasse B/BE</li><li>God evne til kundebehandling</li><li>Er selvstendig og strukturert</li><li>Grunnleggende data kunnskaper.</li><li>Kjennskap til NS 3420, NS 8409 og Bustadoppføringslova.</li></ul><p><br></p><p><strong>Part Construction AB kan tilby:</strong></p><ul><li>Et dynamisk og spennende arbeidsmiljø</li><li>Utfordrende og variert arbeidsoppgaver</li><li>Konkurransedyktig betingelser</li><li>Servicebil, mobil telefon</li><li>Gode forsikringer og pensjons ordninger</li></ul><p><br></p><p><strong>Om arbeidsgiveren </strong></p><p>Part Construction AB, en av Europas ledende produsenter av prefabrikkerte baderoms moduler. Part Construction AB tilbyr entreprenørfirmaer og eiendomsselskaper prefabrikkerte baderom til blant annet, hoteller, sykehjem og boliger. Bademodulene leveres komplett med fliser, gulv, møbler, inventar, helt etter kundens ønsker. rørleggerarbeid og elektriske installasjoner er også ferdig montert fra fabrikk, Part er et familieeid selskap fra Kalix, med ca. 300 ansatte og med en omsetning på ca. 600 millioner. Hovedkvarter og produksjon er i Kalix / Överkalix, Norrbotten. Part Construction AB er en del av konserngruppen PartGroup sammen med søsterselskapene PreBad AB, Altor Industri, Isolamin AB, Space Interior AB, og PCS Modulsystem.</p><p><br></p><p><strong>Sektor</strong></p><p>Privat</p><p><strong>Sted</strong></p><p>Vestbygata 55 2003 Lillestrøm</p><p><strong>Bransje</strong></p><p>Bygg og anlegg,</p><p><strong>Stillingsfunksjon</strong></p><p>Håndverker / Andre montører,</p><p>Håndverker,</p><p>Håndverker / Flislegger, rørlegger, elektriker</p><p> </p><p><strong><u>Søknad</u></strong></p><p><br></p><p><strong>Tiltredelse</strong></p><p>Etter avtale</p><p><strong>Varighet</strong></p><p>Fast, heltid (6 måneders prøvetid)</p><p><strong>Bosted</strong></p><p>Oslo området</p><p><strong>Søknadsfrist</strong></p><p>Snarest, og innen 10.03.2021</p><p><strong>Spørsmål </strong>om stillingen rettes til FIRSTNAME LASTNAME Yttervik, Servicekoordinator Norge. Tel: TELEPHONENO , e-post: EMAIL </p><p>Send søknad merket "Servicetekniker Norge" på e-post til: EMAIL </p><p>Denne rekrutteringen skjer helt i Part Constructions regi, og vi ønsker kun direkte kontakt med personlig søkere. </p><p>Vi gleder oss til å høre fra deg!</p><p id="oppenforalla">Öppen för alla\nVi fokuserar på din kompetens, inte dina övriga förutsättningar. Vi är öppna för att anpassa rollen eller arbetsplatsen efter dina behov.</p>',
'ansokningssattEpost': ' EMAIL ', 'ansokningssattViaAF': False,
'anstallningTyp': {'namn': 'Vanlig anställning', 'varde': 'PFZr_Syz_cUq'}, 'antalPlatser': 1,
'arbetsgivareId': '20610987', 'arbetsgivareNamn': 'Partbyggen i Kalix AB', 'arbetsplatsId': '86440232',
'arbetsplatsNamn': 'Part Construction',
'arbetsplatsadress': {'gatuadress': None, 'kommun': None, 'koordinatPrecision': None, 'lan': None,
'land': {'namn': 'Norge', 'varde': 'QJgN_Zge_BzJ'}, 'latitud': None, 'longitud': None,
'postnr': None, 'postort': None},
'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'ejKravSvenskaEngelska': False,
'informationAnsokningssatt': None, 'ingenErfarenhetKravs': False, 'inkluderande': True,
'kallaTyp': 'VIA_ANNONSERA', 'kontaktpersoner': [
{'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se',
'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}],
'korkort': [{'namn': 'B', 'varde': 'VTK8_WRx_GcM'}],
'lonTyp': {'namn': 'Fast månads- vecko- eller timlön', 'varde': 'oG8G_9cW_nRf'},
'organisationsnummer': '5564003357', 'publiceringsdatum': '2021-01-04 16:18:13',
'referens': 'Servicetekniker Norge', 'sistaAnsokningsdatum': '2021-03-31 23:59:59',
'sistaPubliceringsdatum': '2022-06-14 14:52:55',
'sprak': [{'namn': 'Norska', 'varde': 'pnjj_2JX_Fub', 'vikt': 10}], 'telefonnummer': None,
'tillgangTillEgenBil': False, 'uppdateradTid': 1609773493583,
'varaktighetTyp': {'namn': 'Tillsvidare', 'varde': 'a7uU_j21_mkL'}, 'version': '1.0', 'webbadress': 'partab.nu',
'yrkeserfarenheter': [
{'erfarenhetsniva': {'namn': '1-2 års erfarenhet', 'varde': 'LLnd_5GJ_4ju'}, 'namn': 'VVS-montör',
'varde': 'ZZTi_v6g_4cZ', 'vikt': 5}], 'yrkesroll': {'namn': 'VVS-montör', 'varde': 'ZZTi_v6g_4cZ'}},
{'annonsId': 24458381, 'annonsrubrik': 'Servicetekniker i Oslo sökes',
'annonstext': 'Er du en handy og serviceinnstilt person som ønsker en spennende stilling i en fremoverlent og voksende bedrift?\n\nSom servicetekniker vil du ha ansvar for service og hjemmebesøk for Part Construction AB. Du vil behandle innkommende forespørsler, utføre service oppdrag på byggeprosjekter og reklamasjonsarbeid. Arbeidet foregår hovedsakelig på Østlandet, men vi har også leveranser til andre deler av landet. Noe reising vil derfor måtte påberegnes.\n\nOm deg:\n\n\tEn allsidig praktisk fagarbeider. Du er kanskje flislegger, rørlegger, elektriker eller har en annen relevant utdanning\n\tSnakker og skriver godt Norsk\n\tFørerkort klasse B/BE\n\tGod evne til kundebehandling\n\tEr selvstendig og strukturert\n\tGrunnleggende data kunnskaper.\n\tKjennskap til NS 3420, NS 8409 og Bustadoppføringslova.\n\nPart Construction AB kan tilby:\n\n\tEt dynamisk og spennende arbeidsmiljø\n\tUtfordrende og variert arbeidsoppgaver\n\tKonkurransedyktig betingelser\n\tServicebil, mobil telefon\n\tGode forsikringer og pensjons ordninger\n\nOm arbeidsgiveren \n\nPart Construction AB, en av Europas ledende produsenter av prefabrikkerte baderoms moduler. Part Construction AB tilbyr entreprenørfirmaer og eiendomsselskaper prefabrikkerte baderom til blant annet, hoteller, sykehjem og boliger. Bademodulene leveres komplett med fliser, gulv, møbler, inventar, helt etter kundens ønsker. rørleggerarbeid og elektriske installasjoner er også ferdig montert fra fabrikk, Part er et familieeid selskap fra Kalix, med ca. 300 ansatte og med en omsetning på ca. 600 millioner. Hovedkvarter og produksjon er i Kalix / Överkalix, Norrbotten. Part Construction AB er en del av konserngruppen PartGroup sammen med søsterselskapene PreBad AB, Altor Industri, Isolamin AB, Space Interior AB, og PCS Modulsystem.\n\nSektor\n\nPrivat\n\nSted\n\nVestbygata 55 2003 Lillestrøm\n\nBransje\n\nBygg og anlegg,\n\nStillingsfunksjon\n\nHåndverker / Andre montører,\n\nHåndverker,\n\nHåndverker / Flislegger, rørlegger, elektriker\n\n \n\nSøknad\n\nTiltredelse\n\nEtter avtale\n\nVarighet\n\nFast, heltid (6 måneders prøvetid)\n\nBosted\n\nOslo området\n\nSøknadsfrist\n\nSnarest, og innen 10.03.2021\n\nSpørsmål om stillingen rettes til FIRSTNAME LASTNAME Yttervik, Servicekoordinator Norge. Tel: TELEPHONENO , e-post: EMAIL \n\nSend søknad merket "Servicetekniker Norge" på e-post til: EMAIL \n\nDenne rekrutteringen skjer helt i Part Constructions regi, og vi ønsker kun direkte kontakt med personlig søkere. \n\nVi gleder oss til å høre fra deg!',
'annonstextFormaterad': '<p><strong>Er du en handy og serviceinnstilt person som ønsker en spennende stilling i en fremoverlent og voksende bedrift?</strong></p><p>Som servicetekniker vil du ha ansvar for service og hjemmebesøk for Part Construction AB. Du vil behandle innkommende forespørsler, utføre service oppdrag på byggeprosjekter og reklamasjonsarbeid. Arbeidet foregår hovedsakelig på Østlandet, men vi har også leveranser til andre deler av landet. Noe reising vil derfor måtte påberegnes.</p><p><br></p><p><strong>Om deg:</strong></p><ul><li>En allsidig praktisk fagarbeider. Du er kanskje flislegger, rørlegger, elektriker eller har en annen relevant utdanning</li><li>Snakker og skriver godt Norsk</li><li>Førerkort klasse B/BE</li><li>God evne til kundebehandling</li><li>Er selvstendig og strukturert</li><li>Grunnleggende data kunnskaper.</li><li>Kjennskap til NS 3420, NS 8409 og Bustadoppføringslova.</li></ul><p><br></p><p><strong>Part Construction AB kan tilby:</strong></p><ul><li>Et dynamisk og spennende arbeidsmiljø</li><li>Utfordrende og variert arbeidsoppgaver</li><li>Konkurransedyktig betingelser</li><li>Servicebil, mobil telefon</li><li>Gode forsikringer og pensjons ordninger</li></ul><p><br></p><p><strong>Om arbeidsgiveren </strong></p><p>Part Construction AB, en av Europas ledende produsenter av prefabrikkerte baderoms moduler. Part Construction AB tilbyr entreprenørfirmaer og eiendomsselskaper prefabrikkerte baderom til blant annet, hoteller, sykehjem og boliger. Bademodulene leveres komplett med fliser, gulv, møbler, inventar, helt etter kundens ønsker. rørleggerarbeid og elektriske installasjoner er også ferdig montert fra fabrikk, Part er et familieeid selskap fra Kalix, med ca. 300 ansatte og med en omsetning på ca. 600 millioner. Hovedkvarter og produksjon er i Kalix / Överkalix, Norrbotten. Part Construction AB er en del av konserngruppen PartGroup sammen med søsterselskapene PreBad AB, Altor Industri, Isolamin AB, Space Interior AB, og PCS Modulsystem.</p><p><br></p><p><strong>Sektor</strong></p><p>Privat</p><p><strong>Sted</strong></p><p>Vestbygata 55 2003 Lillestrøm</p><p><strong>Bransje</strong></p><p>Bygg og anlegg,</p><p><strong>Stillingsfunksjon</strong></p><p>Håndverker / Andre montører,</p><p>Håndverker,</p><p>Håndverker / Flislegger, rørlegger, elektriker</p><p> </p><p><strong><u>Søknad</u></strong></p><p><br></p><p><strong>Tiltredelse</strong></p><p>Etter avtale</p><p><strong>Varighet</strong></p><p>Fast, heltid (6 måneders prøvetid)</p><p><strong>Bosted</strong></p><p>Oslo området</p><p><strong>Søknadsfrist</strong></p><p>Snarest, og innen 10.03.2021</p><p><strong>Spørsmål </strong>om stillingen rettes til FIRSTNAME LASTNAME Yttervik, Servicekoordinator Norge. Tel: TELEPHONENO , e-post: EMAIL </p><p>Send søknad merket "Servicetekniker Norge" på e-post til: EMAIL </p><p>Denne rekrutteringen skjer helt i Part Constructions regi, og vi ønsker kun direkte kontakt med personlig søkere. </p><p>Vi gleder oss til å høre fra deg!</p><p id="oppenforalla">Öppen för alla\nVi fokuserar på din kompetens, inte dina övriga förutsättningar. Vi är öppna för att anpassa rollen eller arbetsplatsen efter dina behov.</p>',
'ansokningssattEpost': ' EMAIL ', 'ansokningssattViaAF': False,
'anstallningTyp': {'namn': 'Vanlig anställning', 'varde': 'PFZr_Syz_cUq'}, 'antalPlatser': 1,
'arbetsgivareId': '20610987', 'arbetsgivareNamn': 'Partbyggen i Kalix AB', 'arbetsplatsId': '86440232',
'arbetsplatsNamn': 'Part Construction',
'arbetsplatsadress': {'gatuadress': None, 'kommun': None, 'koordinatPrecision': None, 'lan': None,
'land': {'namn': 'Norge', 'varde': 'QJgN_Zge_BzJ'}, 'latitud': None, 'longitud': None,
'postnr': None, 'postort': None},
'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'ejKravSvenskaEngelska': False,
'informationAnsokningssatt': None, 'ingenErfarenhetKravs': False, 'inkluderande': True,
'kallaTyp': 'VIA_ANNONSERA', 'kontaktpersoner': [
{'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se',
'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}],
'korkort': [{'namn': 'B', 'varde': 'VTK8_WRx_GcM'}],
'lonTyp': {'namn': 'Fast månads- vecko- eller timlön', 'varde': 'oG8G_9cW_nRf'},
'organisationsnummer': '5564003357', 'publiceringsdatum': '2021-01-04 16:18:13',
'referens': 'Servicetekniker Norge', 'sistaAnsokningsdatum': '2021-03-31 23:59:59',
'sistaPubliceringsdatum': '2022-06-14 14:52:55',
'sprak': [{'namn': 'Norska', 'varde': 'pnjj_2JX_Fub', 'vikt': 10}], 'telefonnummer': None,
'tillgangTillEgenBil': False, 'uppdateradTid': 1609773493583,
'varaktighetTyp': {'namn': 'Tillsvidare', 'varde': 'a7uU_j21_mkL'}, 'version': '1.0', 'webbadress': 'partab.nu',
'yrkeserfarenheter': [
{'erfarenhetsniva': {'namn': '1-2 års erfarenhet', 'varde': 'LLnd_5GJ_4ju'}, 'namn': 'VVS-montör',
'varde': 'ZZTi_v6g_4cZ', 'vikt': 5}], 'yrkesroll': {'namn': 'VVS-montör', 'varde': 'ZZTi_v6g_4cZ'}},
{'annonsId': 24513844, 'annonsrubrik': 'Svensktalande Mötesbokare till Soliga Malaga, Spanien!',
'annonstext': 'Svensktalande B2B mötesbokare till soliga Benalmadena i Spanien!\n \nVi söker nu en mötesbokare som kan hjälpa vår partner att ytterligare stärka sin position på företagsmarknaden. Vi söker säljare som är bra med människor, som vill utvecklas vidare som säljare som har förmågan att stå på sig och ge lite extra för att nå resultat. Medarbetare med goda empatiska färdigheter och som kan socialisera och prata med många olika personligheter.\n \nI din roll kommer du att boka möten för vår klient per telefon. Du kommer att arbeta för den svenska marknaden och kontakta verksamheter för att kunna väcka intresse, och sätta upp ett möte för den potentiella nya klienten och säljteamet.\n \nOm dig\nDu kommunicerar väl i tal och skrift samt tycker om att arbeta med uppsökande försäljning mot nya kunder. Vidare drivs du av bra värderingar och arbetar för att alla ska sträva mot samma mål. Likaså har du förmågan att ge det lilla extra för att uppnå dina egna mål. Du har ett starkt driv att uppnå goda resultat och är lösningsorienterad med en god arbetsmoral. Därtill har du hög förmåga att förvärva ny kunskap samt har erfarenhet av och trivs med att ha telefonen som arbetsredskap. \n\nVår klient erbjuder\n-Säljutbildning och vägledning i att arbeta med ett välkänt och starkt varumärke\n-Möjlighet att vidareutveckla dig själv med fokus på kompetensutveckling. \n-En ambitiös och stabil arbetsgivare med långsiktiga mål. \n-Konkurrenskraftig lön. \n-En energisk och ambitiös arbetsmiljö där alla hjälper varandra.\n\nAndra detaljer:\n-Arbetstider: Måndag-fredag 08:00-17:00\n-Heltidsanställning\n-Kontor i Benalmadena, Spanien\n\nKompetenskrav: \n-Flytande svenska både i tal och skrift \n-Flytande engelska både i tal och skrift \n-God teknisk förståelse\n\nFlyttpaket: Återbetald flybjljett + fyra veckors boende vid start samt hjälp med NIE-nummer.\n\nBeskrivning av företaget:\nVår klient är en välkänd aktör på företagsmarknaden - detta är en mycket spännande och utvecklande position. Denna position är placerad i vackra och soliga Benalmadena i Spanien. \n\nVi vill inte bli kontaktade av rekryterings-/bemanningsföretag',
'annonstextFormaterad': 'Svensktalande B2B mötesbokare till soliga Benalmadena i Spanien!\n \nVi söker nu en mötesbokare som kan hjälpa vår partner att ytterligare stärka sin position på företagsmarknaden. Vi söker säljare som är bra med människor, som vill utvecklas vidare som säljare som har förmågan att stå på sig och ge lite extra för att nå resultat. Medarbetare med goda empatiska färdigheter och som kan socialisera och prata med många olika personligheter.\n \nI din roll kommer du att boka möten för vår klient per telefon. Du kommer att arbeta för den svenska marknaden och kontakta verksamheter för att kunna väcka intresse, och sätta upp ett möte för den potentiella nya klienten och säljteamet.\n \nOm dig\nDu kommunicerar väl i tal och skrift samt tycker om att arbeta med uppsökande försäljning mot nya kunder. Vidare drivs du av bra värderingar och arbetar för att alla ska sträva mot samma mål. Likaså har du förmågan att ge det lilla extra för att uppnå dina egna mål. Du har ett starkt driv att uppnå goda resultat och är lösningsorienterad med en god arbetsmoral. Därtill har du hög förmåga att förvärva ny kunskap samt har erfarenhet av och trivs med att ha telefonen som arbetsredskap. \n\nVår klient erbjuder\n-Säljutbildning och vägledning i att arbeta med ett välkänt och starkt varumärke\n-Möjlighet att vidareutveckla dig själv med fokus på kompetensutveckling. \n-En ambitiös och stabil arbetsgivare med långsiktiga mål. \n-Konkurrenskraftig lön. \n-En energisk och ambitiös arbetsmiljö där alla hjälper varandra.\n\nAndra detaljer:\n-Arbetstider: Måndag-fredag 08:00-17:00\n-Heltidsanställning\n-Kontor i Benalmadena, Spanien\n\nKompetenskrav: \n-Flytande svenska både i tal och skrift \n-Flytande engelska både i tal och skrift \n-God teknisk förståelse\n\nFlyttpaket: Återbetald flybjljett + fyra veckors boende vid start samt hjälp med NIE-nummer.\n\nBeskrivning av företaget:\nVår klient är en välkänd aktör på företagsmarknaden - detta är en mycket spännande och utvecklande position. Denna position är placerad i vackra och soliga Benalmadena i Spanien. \n\nVi vill inte bli kontaktade av rekryterings-/bemanningsföretag',
'ansokningssattEpost': ' EMAIL ', 'ansokningssattViaAF': False,
'anstallningTyp': {'namn': 'Arbete utomlands', 'varde': '9Wuo_2Yb_36E'}, 'antalPlatser': 8, 'arbetsgivareId': '0',
'arbetsgivareNamn': 'Nordic Jobs Worldwide AS', 'arbetsplatsId': '21052284',
'arbetsplatsNamn': 'Nordic Jobs Worldwide AS',
'arbetsplatsadress': {'gatuadress': None, 'kommun': None, 'koordinatPrecision': None, 'lan': None,
'land': {'namn': 'Spanien', 'varde': 'bN7k_4ka_YGQ'}, 'latitud': None, 'longitud': None,
'postnr': None, 'postort': None},
'arbetstidTyp': {'namn': 'Heltid', 'varde': '6YE1_gAC_R2G'}, 'avpublicerad': False, 'ejKravSvenskaEngelska': False,
'externtAnnonsId': '0021-045667', 'informationAnsokningssatt': None, 'ingenErfarenhetKravs': True,
'inkluderande': False, 'kallaTyp': 'VIA_AIS', 'kontaktpersoner': [
{'befattning': None, 'beskrivning': 'blabla', 'efternamn': 'Testsson', 'epost': 'test@jobtechdev.se',
'fackligRepresentant': False, 'fornamn': 'Testy', 'telefonnummer': '+01011122233'}],
'lonTyp': {'namn': 'Fast och rörlig lön', 'varde': 'asrX_9Df_ukn'}, 'lonebeskrivning': 'Fast lön + provision',
'publiceringsdatum': '2021-02-01 10:57:26', 'sistaAnsokningsdatum': '2021-04-27 23:59:59',
'sistaPubliceringsdatum': '2022-06-14 14:55:21',
'sprak': [{'namn': 'Svenska', 'varde': 'zSLA_vw2_FXN', 'vikt': 10},
{'namn': 'Engelska', 'varde': 'NVxJ_hLg_TYS', 'vikt': 10}], 'telefonnummer': None,
'tillgangTillEgenBil': False, 'uppdateradTid': 1612173446756,
'varaktighetTyp': {'namn': 'Tillsvidare', 'varde': 'a7uU_j21_mkL'}, 'version': '2.0',
'villkorsbeskrivning': 'Arbetstider: Måndag-fredag 08:00-17:00',
'yrkesroll': {'namn': 'Besöksbokare/Kundbokare', 'varde': 'w8eg_Ufq_B5X'}},
]
ads_no_address = [
{'arbetsplatsadress': {'gatuadress': None, 'kommun': None, 'koordinatPrecision': None, 'lan': None,
'land': {'namn': None, 'varde': None}, 'latitud': None, 'longitud': None,
'postnr': None, 'postort': None}},
] | 177.297619 | 3,199 | 0.738065 | 5,899 | 44,679 | 5.568062 | 0.196983 | 0.004932 | 0.006333 | 0.001827 | 0.899805 | 0.890702 | 0.885526 | 0.87688 | 0.867655 | 0.860592 | 0 | 0.028646 | 0.160075 | 44,679 | 252 | 3,200 | 177.297619 | 0.846617 | 0 | 0 | 0.309623 | 0 | 0.054393 | 0.858259 | 0.116697 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0.016736 | 0.008368 | 0 | 0.008368 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 9 |
faebbcbc5afca41622c242d6955bb7d2f1d7c2ee | 3,174 | py | Python | kachery/parser_helpers.py | flatironinstitute/kachery | d1076f6e8e2df26d3440fdb89f366ec44a502b9b | [
"Apache-2.0"
] | 8 | 2020-03-05T19:41:03.000Z | 2021-11-19T04:40:10.000Z | kachery/parser_helpers.py | flatironinstitute/kachery | d1076f6e8e2df26d3440fdb89f366ec44a502b9b | [
"Apache-2.0"
] | 8 | 2019-11-15T03:40:07.000Z | 2020-09-08T22:14:07.000Z | kachery/parser_helpers.py | flatironinstitute/kachery | d1076f6e8e2df26d3440fdb89f366ec44a502b9b | [
"Apache-2.0"
] | 2 | 2020-08-06T19:56:12.000Z | 2021-09-23T01:05:24.000Z | import os
import kachery as ka
def _add_download_args(parser):
parser.add_argument('--fr', '-f', help='Where to load from', required=False, default='')
parser.add_argument('--remote-only', action='store_true', help='Whether to only load from remote (good for debugging)')
parser.add_argument('--url', help='The URL of the kachery database server to download from when loading from remote (or use KACHERY_URL environment variable)', required=False, default=None)
parser.add_argument('--channel', '-c', help='The channel of the kachery database server to download from when loading from remote (or use KACHERY_CHANNEL environment variable)', required=False, default=None)
parser.add_argument('--password', '-p', help='The password of the kachery database server to download from when loading from remote (or use KACHERY_PASSWORD environment variable)', required=False, default=None)
def _add_upload_args(parser):
parser.add_argument('--to', '-t', help='Where to store to', required=False, default='')
parser.add_argument('--remote-only', action='store_true', help='Whether to only store to remote (good for saving disk space)')
parser.add_argument('--url', help='The URL of the kachery database server to upload to when storing to remote (or use KACHERY_URL environment variable)', required=False, default=None)
parser.add_argument('--channel', '-c', help='The channel of the kachery database server to upload to when storing to remote (or use KACHERY_CHANNEL environment variable)', required=False, default=None)
parser.add_argument('--password', '-p', help='The password of the kachery database server to upload to when storing to remote (or use KACHERY_PASSWORD environment variable)', required=False, default=None)
def _set_download_config_from_parsed_args(args):
fr = args.fr or None
url = args.url or None
channel = args.channel or None
password = args.password or None
remote_only = args.remote_only
if fr is not None:
if url is not None or channel is not None or password is not None:
raise Exception('Cannot use --url or --channel or --password together with --fr')
ka.set_config(
fr=fr,
from_remote_only = remote_only
)
else:
ka.set_config(
fr=dict(
url=url,
channel=channel,
password=password
),
from_remote_only = remote_only
)
def _set_upload_config_from_parsed_args(args):
to = args.to or None
url = args.url or None
channel = args.channel or None
password = args.password or None
remote_only = args.remote_only
if to is not None:
if url is not None or channel is not None or password is not None:
raise Exception('Cannot use --url or --channel or --password together with --to')
ka.set_config(
to=to,
to_remote_only = remote_only
)
else:
ka.set_config(
to=dict(
url=url,
channel=channel,
password=password
),
to_remote_only = remote_only
) | 49.59375 | 214 | 0.669502 | 442 | 3,174 | 4.680995 | 0.147059 | 0.067666 | 0.082165 | 0.057999 | 0.885452 | 0.81102 | 0.81102 | 0.772354 | 0.738521 | 0.738521 | 0 | 0 | 0.23724 | 3,174 | 64 | 215 | 49.59375 | 0.854606 | 0 | 0 | 0.482759 | 0 | 0.103448 | 0.357795 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.068966 | false | 0.172414 | 0.034483 | 0 | 0.103448 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 7 |
879d727eb5dcbb7eb73833b8f03577b34d9910bb | 78,721 | py | Python | src/symbols.py | praga2018/SAMoCAD | 4d1361e38bb2610753a588ca1d5e7391fa4a865e | [
"Apache-2.0"
] | 5 | 2015-05-12T03:58:29.000Z | 2020-05-01T14:20:54.000Z | src/symbols.py | praga2018/SAMoCAD | 4d1361e38bb2610753a588ca1d5e7391fa4a865e | [
"Apache-2.0"
] | 1 | 2019-07-24T01:20:18.000Z | 2019-07-24T01:20:18.000Z | src/symbols.py | sallecta/PyTkCad | 113994559d8d56c02dc102cce53bc7828588fc6c | [
"Apache-2.0"
] | 2 | 2019-03-13T23:59:00.000Z | 2022-01-27T07:04:49.000Z | # -*- coding: utf-8; -*-
import rotate_object
zoomm = 0.8
zoomp = 1.0/0.8
def font(x, y, text, size, zoomOLD, s_s, w_text, anchor, font, angle):
if font == 'Architectural':
tt = Text_arch(x, y, text, size, zoomOLD, s_s, w_text, anchor, angle)
else:
tt = Text_TXT(x, y, text, size, zoomOLD, s_s, w_text, anchor, angle)
return tt
class Text:#Общий класс шрифта
def __init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle):
if not zoomOLD:
self.sy = -size
else:
if zoomOLD>0:
self.sy = -size * (zoomp**zoomOLD)
else:
zoomOLDx=zoomOLD*(-1)
self.sy = -size * (zoomm**zoomOLDx)
self.w_text = w_text #Ширина символа
self.s_s = s_s#смещение символов в строке
self.anchor = anchor#Привязка надписи к базовой точке
self.sx=self.sy/4.0#Ширина символа = четверть высоты
self.liter = -self.s_s#Позиция первого символа
self.Ltext = (len(text) * self.s_s * 2.0 * self.sx) - self.sx #- self.sx#Длина готовой строки
axy = tuple(anchor)
if axy[0] == 's':
self.nachTextY = y
else:# axy[0] == 'n':
self.nachTextY = y + self.sy
if axy[1] == 'w':
self.nachTextX = x
elif axy[1] == 'e':
self.nachTextX = x-self.Ltext
else:# axy[1] == 'c':
self.nachTextX = x-self.Ltext/2.0
self.nabor = []#Список координат символов текста
self.snapLine = [[self.nachTextX,self.nachTextY, self.nachTextX+self.Ltext,self.nachTextY]]#Координаты линии привязки текста
self.nabor.extend(self.snapLine)#Добавить координаты в список
#Дальше работает конкретный шрифт
class Text_arch(Text):#Ахитектурный шрифт
def __init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle):
Text.__init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle)
liter_dict = {
u'А':self.a,
'A':self.a,
'a':self.a,
u'а':self.a,
u'б':self.r6,
u'Б':self.r6,
'6':self.r6,
u'В':self.b,
u'в':self.b,
'B':self.b,
'b':self.b,
'8':self.b,
u'С':self.c,
'C':self.c,
'c':self.c,
u'с':self.c,
u'Г':self.rg,
u'г':self.rg,
u'Д':self.rd,
u'д':self.rd,
u'Е':self.e,
u'е':self.e,
u'Ё':self.e,
u'ё':self.e,
'e':self.e,
'E':self.e,
u'Ж':self.rj,
u'ж':self.rj,
u'З':self.r3,
u'з':self.r3,
'3':self.r3,
u'И':self.ri,
u'и':self.ri,
u'Й':self.rikr,
u'й':self.rikr,
u'К':self.k,
u'к':self.k,
'K':self.k,
'k':self.k,
u'Л':self.rl,
u'л':self.rl,
u'М':self.m,
u'м':self.m,
'M':self.m,
'm':self.m,
u'Н':self.h,
u'н':self.h,
'H':self.h,
'h':self.h,
u'О':self.o,
u'о':self.o,
'O':self.o,
'o':self.o,
'0':self.o,
u'П':self.rp,
u'п':self.rp,
u'Р':self.p,
u'р':self.p,
'P':self.p,
'p':self.p,
u'Т':self.t,
u'т':self.t,
'T':self.t,
't':self.t,
u'У':self.y,
u'у':self.y,
'Y':self.y,
'y':self.y,
u'Ф':self.rf,
u'ф':self.rf,
u'Х':self.x,
u'х':self.x,
'X':self.x,
'x':self.x,
u'Ч':self.rch,
u'ч':self.rch,
u'Ц':self.rc,
u'ц':self.rc,
u'Ш':self.rsh,
u'ш':self.rsh,
u'Щ':self.rcsh,
u'щ':self.rcsh,
u'Ь':self.rmz,
u'ь':self.rmz,
u'Ъ':self.rtz,
u'ъ':self.rtz,
u'Ы':self.rii,
u'ы':self.rii,
u'Э':self.rae,
u'э':self.rae,
u'Ю':self.ru,
u'ю':self.ru,
u'Я':self.rya,
u'я':self.rya,
'D':self.d,
'd':self.d,
'I':self.i,
'i':self.i,
'J':self.j,
'j':self.j,
'F':self.f,
'f':self.f,
'G':self.g,
'g':self.g,
'L':self.L,
'l':self.L,
'N':self.n,
'n':self.n,
'Q':self.q,
'q':self.q,
'R':self.r,
'r':self.r,
'S':self.r5,
's':self.r5,
'U':self.u,
'u':self.u,
'V':self.v,
'v':self.v,
'W':self.w,
'w':self.w,
'Z':self.z,
'z':self.z,
'1':self.r1,
'2':self.r2,
'4':self.r4,
'5':self.r5,
'7':self.r7,
'9':self.r9,
',':self.comma,
'.':self.point,
';':self.c_point,
':':self.p_point,
'!':self.emark,
'?':self.qmark,
'(':self.lbkt,
')':self.rbkt,
'*':self.star,
"/":self.dr,
'+':self.plus,
'-':self.minus,
'=':self.equal,
'>':self.bigest,
'<':self.smollest,
u'№':self.num,
' ':self.space,
}
for i in text:#Перебрать символы строки
self.liter += self.s_s #Передвинуть позицию на один символ
try:
r = liter_dict[i]()
except KeyError:
r = self.qmark()
self.nabor.extend(r)#Добавить в список координаты символа
if angle:
self.nabor = rotate_object.rotate_lines(x, y, self.nabor, angle)
'''
if i in (u'А', 'A', 'a', u'а'):#Если символ А
r = self.a()#Вызвать функцию, возвращающую список координат линий символа А
elif i in (u'б', u'Б', '6'):
r = self.r6()
elif i in (u'В', u'в', 'B', 'b', '8'):
r = self.b()
elif i in (u'С', u'с', 'C', 'c'):
r = self.c()
elif i in (u'Г', u'г'):
r = self.rg()
elif i in (u'Д', u'д'):
r = self.rd()
elif i in (u'Е', u'е', u'Ё', u'ё', 'E', 'e'):
r = self.e()
elif i in (u'Ж', u'ж'):
r = self.rj()
elif i in (u'З', u'з', '3'):
r = self.r3()
elif i in (u'И', u'и'):
r = self.ri()
elif i in (u'Й', u'й'):
r = self.rikr()
elif i in (u'К', u'к', 'K', 'k'):
r = self.k()
elif i in (u'Л', u'л'):
r = self.rl()
elif i in (u'М', u'м', 'M', 'm'):
r = self.m()
elif i in (u'Н', u'н', 'H', 'h'):
r = self.h()
elif i in (u'О', u'о', 'O', 'o', '0'):
r = self.o()
elif i in (u'П', u'п'):
r = self.rp()
elif i in (u'Р', u'р', 'P', 'p'):
r = self.p()
elif i in (u'Т', u'т', 'T', 't'):
r = self.t()
elif i in (u'У', u'у', 'Y', 'y'):
r = self.y()
elif i in (u'Ф', u'ф'):
r = self.rf()
elif i in (u'Х', u'х', 'X', 'x'):
r = self.x()
elif i in (u'Ч', u'ч'):
r = self.rch()
elif i in (u'Ц', u'ц'):
r = self.rc()
elif i in (u'Ш', u'ш'):
r = self.rsh()
elif i in (u'Щ', u'щ'):
r = self.rcsh()
elif i in (u'Ь', u'ь'):
r = self.rmz()
elif i in (u'Ъ', u'ъ'):
r = self.rtz()
elif i in (u'Ы', u'ы'):
r = self.rii()
elif i in (u'Э', u'э'):
r = self.rae()
elif i in (u'Ю', u'ю'):
r = self.ru()
elif i in (u'Я', u'я'):
r = self.rya()
#Английские буквы (не объявленные ранее)
elif i in ('D', 'd'):
r = self.d()
elif i in ('I', 'i'):
r = self.i()
elif i in ('J', 'j'):
r = self.j()
elif i in ('F', 'f'):
r = self.f()
elif i in ('G', 'g'):
r = self.g()
elif i in ('L', 'l'):
r = self.L()
elif i in ('N', 'n'):
r = self.n()
elif i in ('Q', 'q'):
r = self.q()
elif i in ('R', 'r'):
r = self.r()
elif i in ('U', 'u'):
r = self.u()
elif i in ('V', 'v'):
r = self.v()
elif i in ('W', 'w'):
r = self.w()
elif i in ('Z', 'z'):
r = self.z()
#Цифры (которые не объявлены ранее)
elif i == '1':
r = self.r1()
elif i == '2':
r = self.r2()
elif i == '4':
r = self.r4()
elif i in ('5', 's', 'S'):
r = self.r5()
elif i == '7':
r = self.r7()
elif i == '9':
r = self.r9()
#Знаки препинания
elif i == ',':
r = self.comma()
elif i == '.':
r = self.point()
elif i == ';':
r = self.c_point()
elif i == ':':
r = self.p_point()
elif i == '!':
r = self.emark()
elif i == '?':
r = self.qmark()
elif i == '(':
r = self.lbkt()
elif i == ')':
r = self.rbkt()
elif i == '*':
r = self.star()
elif i == "/":
r = self.dr()
elif i == '+':
r = self.plus()
elif i == '-':
r = self.minus()
elif i == '=':
r = self.equal()
elif i == '>':
r = self.bigest()
elif i == '<':
r = self.smollest()
elif i == u'№':
r = self.num()
elif i == ' ':
r = self.space()
else:
r = self.qmark()
'''
def ns(self, w = 1): #Определяет начало рисования символа
w *= self.w_text #Относительная ширина
x=self.nachTextX+self.sx*self.liter*2 #Начало рисования
y=self.nachTextY
y2 = self.sy #Высота символа
x2 = self.sx * w #Ширина символа
return x,y,x2,y2
#Символы узкого архитектурного шрифта
def a(self):#Символ А
x,y,x2,y2 = self.ns()#Определяет нижнюю левую точку символа и его ширину (x2) и высоту (y2)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0.5,-1.0) , #линия / в А (Y координата п рограмме перевернута)
(1.0,0,0.5,-1.0) , #линия \
(0.25,-0.5,0.75,-0.5) , #линия -
]]
def r6(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-0.75),
(0,0,1.0,0),
(0,-0.75,1.0,-0.75),
(0,-1.0,1.0,-1.0),
]]
def b(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,0),
(0,-0.75,1.0,-0.75),
(0,-1.0,1.0,-1.0),
]]
def c(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,0,1.0,0),
(0,-1.0,1.0,-1.0),
]]
def rd(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0.5,-1.0),
(1.0,0,0.5,-1.0),
(-1.0/10.0,0,11.0/10.0,0),
(-1.0/10.0,0,-1.0/10.0,1.0/8.0),
(11.0/10.0,0,11.0/10.0,1.0/8.0),
]]
def rg(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,1.0,-1.0),
]]
def e(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,0,1.0,0),
(0,-0.75,1.0,-0.75),
(0,-1.0,1.0,-1.0),
]]
def r3(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(0,0,1.0,0),
(0,-0.75,1.0,-0.75),
(0,-1.0,1.0,-1.0),
]]
def ri(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-0.75,1.0,-1.0),
]]
def rikr(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-0.75,1.0,-1.0),
(0.5,-16.0/15.0,1.0,-7.0/6.0),
]]
def k(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,-0.25,1.0,-1.0),
(0,-0.25,1.0,-0.25),
(0,-0.25,1.0,0),
]]
def rl(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,1.0/2,-1.0),
(1.0,0,1.0/2,-1.0),
]]
def m(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,0.5,-0.75),
(0.5,-0.75,1.0,-1.0),
]]
def h(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-0.75,1.0,-0.75),
]]
def o(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,0),
(0,-1.0,1.0,-1.0),
]]
def rp(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,1.0,-1.0),
]]
def p(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,-0.25,1.0,-1.0),
(0,-0.25,1.0,-0.25),
(0,-1.0,1.0,-1.0),
]]
def t(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0),
(0,-1.0,1.0,-1.0),
]]
def y(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(0,-0.25,0,-1.0),
(0,-0.25,1.0,-0.25),
(0,0,1.0,0),
]]
def rf(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.25,0,-0.75),
(1.0,-0.25,1.0,-0.75),
(0,-0.25,1.0,-0.25),
(0,-0.75,1.0,-0.75),
(0.5,0,0.5,-1.0),
]]
def x(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,1.0,-1.0),
(0,-1.0,1.0,0),
]]
def rch(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(0,-0.25,0,-1.0),
(0,-0.25,1.0,-0.25),
]]
def rc(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,11.0/10.0,0),
(11.0/10.0,0,11.0/10.0,1/8.0),
]]
def rsh(self):
x,y,x2,y2 = self.ns(w = 1.5)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,0),
(0.5,0,0.5,-1.0),
]]
def rcsh(self):
x,y,x2,y2 = self.ns(w = 1.5)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,11.0/10.0,0),
(11.0/10.0,0,11.0/10.0,1/8.0),
(0.5,0,0.5,-1.0),
]]
def rmz(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-0.75),
(0,0,1.0,0),
(0,-0.75,1.0,-0.75),
]]
def rtz(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-0.75),
(0,0,1.0,0),
(0,-0.75,1.0,-0.75),
(0,-1.0,-1.0/6.0,-1.0),
]]
def rii(self):
x,y,x2,y2 = self.ns(1.5)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,0.66,0),
(0,-0.75,0.66,-0.75),
(0.66,0,0.66,-0.75),
]]
def rae(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(0,0,1.0,0),
(0,-0.5,1.0,-0.5),
(0,-1.0,1.0,-1.0),
]]
def ru(self):
x,y,x2,y2 = self.ns(1.5)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0.33,0,0.33,-1.0),
(0.33,0,1.0,0),
(0.33,-1.0,1.0,-1.0),
(0,-0.5,0.33,-0.5),
]]
def rya(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(0,-0.25,0,-1.0),
(0,-0.25,1.0,-0.25),
(0,0,1.0,-0.25),
(0,-1.0,1.0,-1.0),
]]
def rj(self):
x,y,x2,y2 = self.ns(w = 1.5)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.25,0,-1.0),
(1.0,-0.25,1.0,-1.0),
(0,-0.25,1.0,-0.25),
(0.5,0,0.5,-1.0),
(0.5,-0.25,0,0),
(0.5,-0.25,1.0,0),
]]
#Английские буквы (не объявленные ранее)
def d(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,0,0.66,0),
(0,-1.0,0.66,-1.0),
(0.66,0,1.0,-0.25),
(0.66,-1.0,1.0,-0.75),
(1.0,-0.75,1.0,-0.25),
]]
def i(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0),
(0,-1.0,1.0,-1.0),
(0,0,1.0,0),
]]
def j(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0),
(0,-1.0,1.0,-1.0),
(0,0,0.5,0),
(0,0,0,-1.0/5.0),
]]
def f(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-0.75,1.0,-0.75),
(0,-1.0,1.0,-1.0),
]]
def g(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,0,1.0,0),
(0,-1.0,1.0,-1.0),
(1.0,0,1.0,-1.0/6.0),
(7.0/6.0,-1.0/6.0,5.0/6.0,-1.0/6.0),
]]
def L(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,0,1.0,0),
]]
def n(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,1.0,0),
]]
def q(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,0),
(0,-1.0,1.0,-1.0),
(4.0/5.0,-1.0/10.0,6.0/5.0,1/10.0),
]]
def r(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,-0.25,1.0,-1.0),
(0,-0.25,1.0,-0.25),
(0,-1.0,1.0,-1.0),
(1.0,0,0,-0.25),
]]
def u(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,0),
]]
def v(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.5,0),
(1.0,-1.0,0.5,0),
]]
def w(self):
x,y,x2,y2 = self.ns(1.5)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.25,0),
(0.5,-1.0,0.25,0),
(0.5,-1.0,0.75,0),
(0.75,0,1.0,-1.0),
]]
def z(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-1.0,0,0),
(0,-1.0,1.0,-1.0),
(0,0,1.0,0),
]]
#Цифры (не заданные ранее)
def r1(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(0,-0.75,1.0,-1.0),
]]
def r2(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-7.0/8.0,0,0),
(1.0,-1.0,1.0,-7.0/8.0),
(0,-7.0/8.0,0,-1.0),
(0,-1.0,1.0,-1.0),
(0,0,1.0,0),
]]
def r4(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(1.0,-1.0,0,-0.25),
(0,-0.25,1.0,-0.25),
]]
def r5(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.75,0,-1.0),
(1.0,0,1.0,-0.75),
(0,0,1.0,0),
(0,-0.75,1.0,-0.75),
(0,-1.0,1.0,-1.0),
]]
def r7(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-1.0,0,0),
(0,-7.0/8.0,0,-1.0),
(0,-1.0,1.0,-1.0),
]]
def r9(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(0,-0.25,0,-1.0),
(0,-0.25,1.0,-0.25),
(0,0,1.0,0),
(0,-1.0,1.0,-1.0),
]]
#Знаки
def comma(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,-0.33,1/5.0),
]]
def point(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,-1.0/16.0,0),
]]
def c_point(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,-0.33,1/5.0),
(0,-1.0,-1.0/16.0,-1.0),
]]
def p_point(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,-1.0/16.0,0),
(0,-1.0,-1.0/16.0,-1.0),
]]
def emark(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,-1.0/16.0,0),
(-1.0/32.0,-1.0/10.0,-1.0/32.0,-1.0),
]]
def qmark(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,-1.0/16.0,0),
(1.0,-7.0/8.0,0,-0.25),
(0,-0.25,0,-1.0/8.0),
(1.0,-1.0,1.0,-7.0/8.0),
(0,-7.0/8.0,0,-1.0),
(0,-1.0,1.0,-1.0),
]]
def lbkt(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,0,-1.0/4),
(1.0,-1.0,0,-0.75),
(0,-0.75,0,-0.25),
]]
def rbkt(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,1.0,-0.25),
(0,-1.0,1.0,-0.75),
(1.0,-0.75,1.0,-0.25),
]]
def star(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,1.0,-0.75),
(1.0,-1.0,0,-0.75),
(0.5,-1.0,0.5,-0.75),
]]
def dr(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-1.0,0,0),
]]
def plus(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,-0.66,0.5,-0.33),
(0,-0.5,1.0,-0.5),
]]
def minus(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.5,1.0,-0.5),
]]
def equal(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,1.0,-0.33),
(0,-0.5,1.0,-0.5),
]]
def bigest(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0/3.5,1.0,-0.33),
(0,-1.0/3.5,1.0,-0.25),
]]
def smollest(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,1.0,-1.0/3.5),
(0,-0.25,1.0,-1.0/3.5),
]]
def num(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,1.0,0),
(7.0/6.0,-1.0,8.0/6.0,-1.0),
(1.33,-1.0,1.33,-0.75),
(1.33,-0.75,7.0/6.0,-0.75),
(7.0/6.0,-0.75,7.0/6.0,-1.0),
(1.33,-0.5,7.0/6.0,-0.5),
]]
def space(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
]]
class Text_TXT(Text):
def __init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle):
Text.__init__(self, x, y, text, size, zoomOLD, s_s, w_text, anchor, angle)
liter_dict = {
u'А':self.a,
'A':self.a,
'a':self.am,
u'а':self.am,
u'б':self.rbm,
u'Б':self.rb,
u'В':self.b,
u'в':self.rvm,
'B':self.b,
'b':self.bm,
u'С':self.c,
'C':self.c,
'c':self.cm,
u'с':self.cm,
u'Г':self.rg,
u'г':self.rgm,
u'Д':self.rd,
u'д':self.rdm,
u'Е':self.e,
u'е':self.em,
u'Ё':self.e,
u'ё':self.em,
'e':self.em,
'E':self.e,
u'Ж':self.rj,
u'ж':self.rjm,
u'З':self.rz,
u'з':self.rzm,
u'И':self.ri,
u'и':self.rim,
u'Й':self.rikr,
u'й':self.rikrm,
u'К':self.k,
u'к':self.rkm,
'K':self.k,
'k':self.km,
u'Л':self.rl,
u'л':self.rlm,
u'М':self.m,
u'м':self.rmm,
'M':self.m,
'm':self.mm,
u'Н':self.h,
u'н':self.rnm,
'H':self.h,
'h':self.hm,
u'О':self.ro,
u'о':self.om,
'O':self.o,
'o':self.om,
u'П':self.rp,
u'п':self.rpm,
u'Р':self.p,
u'р':self.pm,
'P':self.p,
'p':self.pm,
u'Т':self.t,
u'т':self.rtm,
'T':self.t,
't':self.tm,
u'У':self.ru,
u'у':self.rum,
u'Ф':self.rf,
u'ф':self.rfm,
u'Х':self.x,
u'х':self.xm,
'X':self.x,
'x':self.xm,
u'Ч':self.rtch,
u'ч':self.rtchm,
u'Ц':self.rtc,
u'ц':self.rtcm,
u'Ш':self.rsh,
u'ш':self.rshm,
u'Щ':self.rtsh,
u'щ':self.rtshm,
u'Ь':self.rmznak,
u'ь':self.rmznakm,
u'Ъ':self.rtznak,
u'ъ':self.rtznakm,
u'Ы':self.rii,
u'ы':self.riim,
u'Э':self.rae,
u'э':self.raem,
u'Ю':self.ryu,
u'ю':self.ryum,
u'Я':self.rya,
u'я':self.ryam,
'D':self.d,
'd':self.dm,
'I':self.i,
'i':self.im,
'J':self.j,
'j':self.jm,
'K':self.k,
'k':self.km,
'F':self.f,
'f':self.fm,
'G':self.g,
'g':self.gm,
'L':self.l,
'l':self.lm,
'N':self.n,
'n':self.nm,
'Q':self.q,
'q':self.qm,
'R':self.r,
'r':self.rm,
'S':self.s,
's':self.sm,
'U':self.u,
'u':self.um,
'V':self.v,
'v':self.vm,
'W':self.w,
'w':self.wm,
'Y':self.y,
'y':self.ym,
'Z':self.z,
'z':self.zm,
'1':self.n1,
'2':self.n2,
'3':self.rz,
'4':self.n4,
'5':self.n5,
'6':self.n6,
'7':self.n7,
'8':self.n8,
'9':self.n9,
'0':self.n0,
',':self.comma,
'.':self.point,
';':self.c_point,
':':self.p_point,
'!':self.emark,
'?':self.qmark,
'(':self.lbkt,
')':self.rbkt,
'*':self.star,
"/":self.dr,
'+':self.plus,
'-':self.minus,
'=':self.equal,
'>':self.bigest,
'<':self.smollest,
u'№':self.num,
' ':self.space,
}
for i in text:
self.liter+=self.s_s
try:
r = liter_dict[i]()
except KeyError:
r = self.qmark()
self.nabor.extend(r)
'''
if i in ('A', u'А'):
r = self.a()
elif i in ('a', u'а'):
r = self.am()
elif i == u'Б':
r = self.rb()
elif i == u'б':
r = self.rbm()
elif i in ('B', u'В'):
r = self.b()
elif i =='b':
r = self.bm()
elif i in ('C', u'С'):
r = self.c()
elif i in ('c', u'с'):
r = self.cm()
elif i =='D':
r = self.d()
elif i =='d':
r = self.dm()
elif i in ('E', u'Е'):
r = self.e()
elif i in ('e', u'е'):
r = self.em()
elif i =='F':
r = self.f()
elif i =='f':
r = self.fm()
elif i =='G':
r = self.g()
elif i =='g':
r = self.gm()
elif i in ('H', u'Н'):
r = self.h()
elif i =='h':
r = self.hm()
elif i =='I':
r = self.i()
elif i =='i':
r = self.im()
elif i =='J':
r = self.j()
elif i =='j':
r = self.jm()
elif i in ('K', u'К'):
r = self.k()
elif i =='k':
r = self.km()
elif i =='L':
r = self.l()
elif i =='l':
r = self.lm()
elif i in ('M', u'М'):
r = self.m()
elif i =='m':
r = self.mm()
elif i =='N':
r = self.n()
elif i =='n':
r = self.nm()
elif i =='O':
r = self.o()
elif i in ('o', u'о'):
r = self.om()
elif i in ('P', u'Р'):
r = self.p()
elif i in ('p', u'р'):
r = self.pm()
elif i =='Q':
r = self.q()
elif i =='q':
r = self.qm()
elif i =='R':
r = self.r()
elif i =='r':
r = self.rm()
elif i =='S':
r = self.s()
elif i =='s':
r = self.sm()
elif i in ('T', u'Т'):
r = self.t()
elif i =='t':
r = self.tm()
elif i =='U':
r = self.u()
elif i =='u':
r = self.um()
elif i =='V':
r = self.v()
elif i =='v':
r = self.vm()
elif i =='W':
r = self.w()
elif i =='w':
r = self.wm()
elif i in ('X', u'Х'):
r = self.x()
elif i in ('x', u'х'):
r = self.xm()
elif i =='Y':
r = self.y()
elif i =='y':
r = self.ym()
elif i =='Z':
r = self.z()
elif i =='z':
r = self.zm()
elif i == u'в':
r = self.rvm()
elif i ==u'Г':
r = self.rg()
elif i ==u'г':
r = self.rgm()
elif i ==u'Д':
r = self.rd()
elif i ==u'д':
r = self.rdm()
elif i ==u'Ж':
r = self.rj()
elif i ==u'ж':
r = self.rjm()
elif i in (u'З', '3'):
r = self.rz()
elif i == u'з':
r = self.rzm()
elif i ==u'И':
r = self.ri()
elif i ==u'и':
r = self.rim()
elif i ==u'Й':
r = self.rikr()
elif i ==u'й':
r = self.rikrm()
elif i ==u'к':
r = self.rkm()
elif i ==u'Л':
r = self.rl()
elif i ==u'л':
r = self.rlm()
elif i ==u'м':
r = self.rmm()
elif i ==u'н':
r = self.rnm()
elif i ==u'П':
r = self.rp()
elif i ==u'п':
r = self.rpm()
elif i ==u'О':
r = self.ro()
elif i ==u'т':
r = self.rtm()
elif i ==u'У':
r = self.ru()
elif i ==u'у':
r = self.rum()
elif i ==u'Ф':
r = self.rf()
elif i ==u'ф':
r = self.rfm()
elif i ==u'Ц':
r = self.rtc()
elif i ==u'ц':
r = self.rtcm()
elif i ==u'Ч':
r = self.rtch()
elif i ==u'ч':
r = self.rtchm()
elif i ==u'Ш':
r = self.rsh()
elif i ==u'ш':
r = self.rshm()
elif i ==u'Щ':
r = self.rtsh()
elif i ==u'щ':
r = self.rtshm()
elif i ==u'Ъ':
r = self.rtznak()
elif i ==u'ъ':
r = self.rtznakm()
elif i ==u'Ы':
r = self.rii()
elif i ==u'ы':
r = self.riim()
elif i ==u'Ь':
r = self.rmznak()
elif i ==u'ь':
r = self.rmznakm()
elif i ==u'Э':
r = self.rae()
elif i ==u'э':
r = self.raem()
elif i ==u'Ю':
r = self.ryu()
elif i ==u'ю':
r = self.ryum()
elif i ==u'Я':
r = self.rya()
elif i ==u'я':
r = self.ryam()
elif i =='1':
r = self.n1()
elif i =='2':
r = self.n2()
elif i =='4':
r = self.n4()
elif i =='5':
r = self.n5()
elif i =='6':
r = self.n6()
elif i =='7':
r = self.n7()
elif i =='8':
r = self.n8()
elif i =='9':
r = self.n9()
elif i =='0':
r = self.n0()
elif i == ' ':
r = self.space()
#Знаки препинания
elif i == ',':
r = self.comma()
elif i == '.':
r = self.point()
elif i == ';':
r = self.c_point()
elif i == ':':
r = self.p_point()
elif i == '!':
r = self.emark()
elif i == '?':
r = self.qmark()
elif i == '(':
r = self.lbkt()
elif i == ')':
r = self.rbkt()
elif i == '*':
r = self.star()
elif i == "/":
r = self.dr()
elif i == '+':
r = self.plus()
elif i == '-':
r = self.minus()
elif i == '=':
r = self.equal()
elif i == '>':
r = self.bigest()
elif i == '<':
r = self.smollest()
elif i == u'№':
r = self.num()
else:
r = self.qmark()
'''
if angle:
self.nabor = rotate_object.rotate_lines(x, y, self.nabor, angle)
def ns(self, w = 1): #Определяет начало рисования символа
w *= self.w_text
x=self.nachTextX+self.sx*self.liter*2 #Начало рисования
y=self.nachTextY
y2 = self.sy
x2 = self.sx * w
return x,y,x2,y2
def a(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-0.33),
(1.0,0,1.0,-0.33),
(0,-0.33,1.0,-0.33),
(0,-0.33,0.5,-1.0),
(1.0,-0.33,0.5,-1.0),
]]
def am(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,0,1.0,-1.0),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def b(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,0.66,-1.0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-4.0/6.0),
(1.0,-4.0/6.0,0.66,-0.5),
]]
def bm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-0.33,1.0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(0,0,0,-1.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def c(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
((1/3.0)*2.0,0,1.0,-0.33),
]]
def cm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,1.0,-1.0),
(0.33,0,1.0,0),
(0.33,0,0,-0.33),
]]
def d(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-0.33,1.0,-0.66),
(-0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(0,0,0,-1.0),
(-0.33,0,0.66,0),
(0.66,0,1.0,-0.33),
]]
def dm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,0,1.0,-1.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def e(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,1.0,-1.0),
(0,0,1.0,0),
(0,-0.5,0.5,-0.5),
]]
def em(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.5,0.66,-0.5),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.75),
(0,-0.33,0,-0.66),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,-0.5,1.0,-0.75),
]]
def f(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,1.0,-1.0),
(0,-0.5,0.5,-0.5),
]]
def fm(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(-0.33,-0.5,0.66,-0.5),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.75),
(0,0,0,-0.66),
]]
def g(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,1.0,-1.0),
(0.33,0,1.0,0),
(0.33,0,0,-0.33),
(1.0,0,1.0,-0.5),
(1.0,-0.5,0.75,-0.5),
]]
def gm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,0.33,1.0,-0.66),
(0.33,0,1.0,0),
(0.33,0,0,-0.33),
(0.66,0.66,1.0,0.33),
(0.33,0.66,0.66,0.66),
(0.33,0.66,0,0.33),
]]
def h(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-0.75,1.0,-0.75),
]]
def hm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(0,0,0,-1.33),
(1.0,0,1.0,-0.66),
]]
def i(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.33,0,0.66,0),
(0.33,-1.0,0.66,-1.0),
(0.5,0,0.5,-1.0),
]]
def im(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0),
(0.5,-1.33,0.5,-3.5/3.0),
]]
def j(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-0.33,1.0,-1.0),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def jm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-0.33,1.0,-1.0),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
(1.0,-1.33,1.0,-3.5/3.0),
]]
def k(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-0.5,0.33,-0.5),
(0.33,-0.5,1.0,-1.0),
(0.33,-0.5,1.0,0),
]]
def km(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.33),
(0,-0.5,0.33,-0.5),
(0.33,-0.5,1.0,-1.0),
(0.33,-0.5,1.0,0),
]]
def l(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,0,1.0,0),
]]
def lm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,-0.25,0.5,-1.0),
(0.5,-0.25,0.75,0),
]]
def m(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,0.5,-0.5),
(0.5,-0.5,1.0,-1.0),
]]
def mm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.75,0.25,-1.0),
(0.25,-1.0,0.5,-0.75),
(0.5,-0.75,0.75,-1.0),
(0.75,-1.0,1.0,-0.75),
(0,0,0,-1.0),
(1.0,0,1.0,-0.75),
(0.5,-0.75,0.5,-0.5),
]]
def n(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,1.0,0),
]]
def nm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(0,0,0,-1.0),
(1.0,0,1.0,-0.66),
]]
def o(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,0),
(0,-1.0,1.0,-1.0),
]]
def om(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,-0.66,1.0,-0.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def p(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,0.66,-1.0),
(0,-0.5,0.66,-0.5),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-4.0/6.0),
(1.0,-4.0/6.0,0.66,-0.5),
]]
def pm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-0.33,1.0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(0,0.5,0,-1.0),
(0,0,0.66,0),
(0.66,0,1.0,-0.33),
]]
def q(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,-0.66,1.0,-0.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
(0.66,-0.33,1.0,0),
]]
def qm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,0.66,1.0,-1.0),
(0.33,0,1.0,0),
(0.33,0,0,-0.33),
]]
def r(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,0.66,-1.0),
(0,-0.5,0.66,-0.5),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-4.0/6.0),
(1.0,-4.0/6.0,0.66,-0.5),
(0.33,-0.5,1.0,0),
]]
def rm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(0,0,0,-1.0),
]]
def s(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-5.0/6.0,1.0/6.0,-1.0),
(1.0/6.0,-1.0,5.0/6.0,-1.0),
(5.0/6.0,-1.0,1.0,-5.0/6.0),
(0,-5.0/6.0,1.0,-1.0/6.0),
(1.0/6.0,0,5.0/6.0,0),
(1.0/6.0,0,0,-1.0/6.0),
(5.0/6.0,0,1.0,-1.0/6.0),
]]
def sm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.75,0.33,-1.0),
(0.33,-1.0,1.0,-1.0),
(0,-0.75,0.33,-0.5),
(0.33,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.25),
(1.0,-0.25,0.66,0),
(0.66,0,0,0),
]]
def t(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0/2,0,1.0/2,-1.0),
(0,-1.0,1.0,-1.0),
]]
def tm(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,-0.25,0.5,-1.0),
(0,-0.66,1.0,-0.66),
(0.5,-0.25,0.75,0),
(0.75,0,1.0,-0.25),
]]
def u(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-1.0),
(1.0,-0.33,1.0,-1.0),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def um(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-1.0),
(0,-0.33,0.33,0),
(0.33,0,0.66,0),
(1.0,-0.33,0.66,0),
(1.0,0,1.0,-1.0),
]]
def v(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.5,0),
(0.5,0,1.0, -1),
]]
def vm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.5,0),
(0.5,0,1.0, -1.0),
]]
def w(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.25,0),
(0.25,0,0.5, -1/2.0),
(0.5, -1/2.0,0.75, 0),
(0.75, 0,1.0, -1),
]]
def wm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.25,0),
(0.25,0,0.5, -1.0),
(0.5, -1.0,0.75, 0),
(0.75, 0,1.0, -1.0),
]]
def x(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,1.0,0),
(0,0,1.0,-1.0),
]]
def xm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,1.0,0),
(0,0,1.0,-1.0),
]]
def y(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.5,-0.5),
(1.0, -1,0.5,-0.5),
(0.5,-0.5,0.5,0),
]]
def ym(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.5,-0.5),
(0,0,1.0,-1.0),
]]
def z(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-1.0,0,0),
(0,-1.0,1.0,-1.0),
(0,0,1.0,0),
]]
def zm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-1.0,0,0),
(0,-1.0,1.0,-1.0),
(0,0,1.0,0),
]]
#Русские буквы (которые не встречались ранее)
def rb(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,0.66,-1.0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
(0.66,-1.0,0.66,-5.0/6.0),
]]
def rbm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,0.66,-1.0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
]]
def rvm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,0.66,-1.0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-4.0/6.0),
(1.0,-4.0/6.0,0.66,-0.5),
]]
def rg(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,1.0,-1.0),
(1.0,-1.0,1.0,-5.0/6.0),
]]
def rgm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,1.0,-1.0),
]]
def rd(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(-1.0/6.0,0,7.0/6.0,0),
(-1.0/6.0,0,-1.0/6.0,1/6.0),
(7.0/6.0,1/6.0,7.0/6.0,0),
(0,0,0,-0.5),
(0,-0.5,0.33,-1.0),
(0.33,-1.0,1.0,-1.0),
(1.0,-1.0,1.0,0),
]]
def rdm(self):
x,y,x2,y2 = self.ns(0.8)
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(-1.0/6.0,0,7.0/6.0,0),
(-1.0/6.0,0,-1.0/6.0,1.0/6.0),
(7.0/6.0,1.0/6.0,7.0/6.0,0),
(0,0,0,-0.5),
(0,-0.5,0.33,-1.0),
(0.33,-1.0,1.0,-1.0),
(1.0,-1.0,1.0,0),
]]
def rj(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,1.0,-1.0),
(0,-1.0,1.0,0),
(0.5,0,0.5,-1.0),
]]
def rjm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,1.0,-1.0),
(0,-1.0,1.0,0),
(0.5,0,0.5,-1.0),
]]
def rz(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,-0.5,0.66,-0.5),
(0.33,-1.0,0.66,-1.0),
(0.33,0,0.66,0),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-4.0/6.0),
(1.0,-4.0/6.0,0.66,-0.5),
(0.33,-1.0,0,-5.0/6.0),
(0.33,0,0,-1.0/6.0),
]]
def rzm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,-0.5,0.66,-0.5),
(0.33,-1.0,0.66,-1.0),
(0.33,0,0.66,0),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-4.0/6.0),
(1.0,-4.0/6.0,0.66,-0.5),
(0.33,-1.0,0,-5.0/6.0),
(0.33,0,0,-1.0/6.0),
]]
def ri(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,-0.66),
]]
def rim(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,-0.66),
]]
def rikr(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,-0.66),
(0.33,-1.0,0.66,-1.0),
]]
def rikrm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,1.0,-0.66),
(0.33,-1.0,0.66,-1.0),
]]
def rkm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-0.5,0.33,-0.5),
(0.33,-0.5,1.0,-1.0),
(0.33,-0.5,1.0,0),
]]
def rl(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.25,0,0.75,-1.0),
(0.75,-1.0,1.0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,0.25,0),
]]
def rlm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0.5,-1.0),
(0.5,-1.0,1.0,-1.0),
(1.0,0,1.0,-1.0),
]]
def rmm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,0.5,-0.5),
(0.5,-0.5,1.0,-1.0),
]]
def rnm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.+0.25,1.0,-1.+0.25),
]]
def ro(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,-0.66,1.0,-0.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def rp(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,1.0,-1.0),
]]
def rpm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,1.0,-1.0),
]]
def rtm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0/2,0,1.0/2,-1.0),
(0,-1.0,1.0,-1.0),
]]
def ru(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0,-0.66),
(0,-0.66,0.25,-0.5),
(1.0,-1.0,1.0,-0.25),
(1.0,-0.25,0.75,0),
(0.75,0,0.25,0),
(0.25,0,0,-0.25),
(0.25,-0.5,1.0,-0.5),
]]
def rum(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0.5,-0.5),
(1.0,-1.0,1.0,-0.25),
(1.0,-0.25,0.75,0),
(0.75,0,0.25,0),
(0.25,0,0,-0.25),
(0.5,-0.5,1.0,-0.5),
]]
def rf(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,-1.0,0.5,0),
(0.25,-1.0,0.75,-1.0),
(0.75,-1.0,1.0,-0.75),
(1.0,-0.75,1.0,-0.5),
(1.0,-0.5,0.75,-0.25),
(0.75,-0.25,0.25,-0.25),
(0.25,-0.25,0,-0.5),
(0,-0.5,0,-0.75),
(0,-0.75,0.25,-1.0),
]]
def rfm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,-1.0,0.5,0),
(0.25,-1.0,0.75,-1.0),
(0.75,-1.0,1.0,-0.75),
(1.0,-0.75,1.0,-0.5),
(1.0,-0.5,0.75,-0.25),
(0.75,-0.25,0.25,-0.25),
(0.25,-0.25,0,-0.5),
(0,-0.5,0,-0.75),
(0,-0.75,0.25,-1.0),
]]
def rtc(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,7.0/6.0,0),
(7.0/6.0,0,7.0/6.0,1.0/6.0),
]]
def rtcm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,0,7.0/6.0,0),
(7.0/6.0,0,7.0/6.0,1.0/6.0),
]]
def rtch(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0,-0.66),
(0,-0.66,0.25,-0.5),
(1.0,0,1.0,-1.0),
(0.25,-0.5,1.0,-0.5),
]]
def rtchm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,0,-0.5),
(1.0,0,1.0,-1.0),
(0,-0.5,1.0,-0.5),
]]
def rsh(self):
x,y,x2,y2 = self.ns(1.2)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0.5,-0.5,0.5,0),
(0,0,1.0,0),
]]
def rshm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0.5,-0.5,0.5,0),
(0,0,1.0,0),
]]
def rtsh(self):
x,y,x2,y2 = self.ns(1.2)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0.5,-0.5,0.5,0),
(0,0,7.0/6.0,0),
(7.0/6.0,0,7.0/6.0,1.0/6.0),
]]
def rtshm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0.5,-0.5,0.5,0),
(0,0,7.0/6.0,0),
(7.0/6.0,0,7.0/6.0,1.0/6.0),
]]
def rtznak(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,-1.0/6.0,-1.0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
(-1.0/6.0,-1.0,-1.0/6.0,-5.0/6.0),
]]
def rtznakm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,-1.0,-1.0/6.0,-1.0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
(-1.0/6.0,-1.0,-1.0/6.0,-5.0/6.0),
]]
def rii(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(7.0/6.0,-1.0,7.0/6.0,0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
]]
def riim(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(7.0/6.0,-1.0,7.0/6.0,0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
]]
def rmznak(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
]]
def rmznakm(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(0,0,0.66,0),
(0,-0.5,0.66,-0.5),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-1.0/6.0),
(1.0,-1.0/6.0,0.66,0),
]]
def rae(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.33,-0.5,1.0,-0.5),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,-0.66,1.0,-0.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def raem(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.33,-0.5,1.0,-0.5),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,-0.66,1.0,-0.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
]]
def ryu(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,-0.66,1.0,-0.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
(-1.0/6.0,-0.5,0,-0.5),
(-1.0/6.0,0,-1.0/6.0,-1.0),
]]
def ryum(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,0,-0.66),
(0,-0.66,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-0.66),
(1.0,-0.66,1.0,-0.33),
(0.33,0,0.66,0),
(0.33,0,0,-0.33),
(0.66,0,1.0,-0.33),
(-1.0/6.0,-0.5,0,-0.5),
(-1.0/6.0,0,-1.0/6.0,-1.0),
]]
def rya(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(1.0,-1.0,0.33,-1.0),
(0.33,-1.0,0,-0.75),
(0,-0.75,0,-0.5),
(0,-0.5,0.25, -1/4.0),
(0.25, -1/4.0,1.0, -1/4.0),
(0.5, -1/4.0,0, 0),
]]
def ryam(self):
x,y,x2,y2 = self.ns()
y2 *= 0.75
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,1.0,-1.0),
(1.0,-1.0,0.33,-1.0),
(0.33,-1.0,0,-0.75),
(0,-0.75,0,-0.5),
(0,-0.5,0.25, -0.25),
(0.25, -0.25,1.0, -0.25),
(0.5, -0.25,0, 0),
]]
#Цифры
def n1(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.33,0,0.66,0),
(0.5,-1.0,0.33,-5.0/6.0),
(0.5,0,0.5,-1.0),
]]
def n2(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-5.0/6.0,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-0.66),
(1.0,-0.66,0.66,-0.5),
(0.66,-0.5,0.33,-0.5),
(0.33,-0.5,0,-0.33),
(0,-0.33,0,0),
(0,0,1.0,0),
]]
def n4(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-1.0,1.0,0),
(1.0,-1.0,0,-0.5),
(0,-0.5,7.0/6.0,-0.5),
]]
def n5(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-1.0,0,-1.0),
(0,-1.0,0,-0.75),
(0,-0.75,0.66,-0.75),
(0.66,-0.75,1.0,-0.5),
(1.0,-0.5,1.0,-0.25),
(1.0,-0.25,0.66,0),
(0.66,0,0.33,0),
(0.33,0,0,-0.25),
]]
def n6(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.66,-1.0,0.33,-1.0),
(0.33,-1.0,0,-0.75),
(0,-0.75,0,-0.25),
(0.66,-0.5,1.0,-0.33),
(1.0,-0.33,1.0,-0.25),
(1.0,-0.25,0.66,0),
(0.66,0,0.33,0),
(0.33,0,0,-0.25),
(0,-0.5,0.66,-0.5),
]]
def n7(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0,1.0,-1.0),
(1.0,-1.0,0,0),
]]
def n8(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-5.0/6.0,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-0.66),
(1.0,-0.66,0.66,-0.5),
(0.66,-0.5,0.33,-0.5),
(0.33,-0.5,0,-0.33),
(0,-0.33,0,-1.0/6.0),
(0,-1.0/6.0,0.33,0),
(0.33,0,0.66,0),
(0.66,0,1.0,-1.0/6.0),
(1.0,-1.0/6.0,1.0,-0.33),
(1.0,-0.33,0.66,-0.5),
(0,-5.0/6.0,0,-0.66),
(0,-0.66,0.33,-0.5),
]]
def n9(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-5.0/6.0,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-1.0/6.0),
(1.0,-0.5,0.33,-0.5),
(0.33,0,0.66,0),
(0.66,0,1.0,-1.0/6.0),
(0,-5.0/6.0,0,-0.66),
(0,-0.66,0.33,-0.5),
]]
def n0(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-1.0/6.0,0,-5.0/6.0),
(0,-5.0/6.0,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,1.0,-1.0/6.0),
(0.33,0,0.66,0),
(0.33,0,0,-1.0/6.0),
(0.66,0,1.0,-1.0/6.0),
]]
#Знаки препинания
def point(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0/6.0),
]]
def comma(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0/6.0),
(0.5,0,0,0.33),
]]
def c_point(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0/6.0),
(0.5,0,0,0.33),
(0.5,-0.33,0.5,-0.5),
]]
def p_point(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0/6.0),
(0.5,-0.33,0.5,-0.5),
]]
def emark(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,0,0.5,-1.0/6.0),
(0.5,-0.33,0.5,-1.0),
]]
def space(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
]]
def qmark(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-5.0/6.0,0.33,-1.0),
(0.33,-1.0,0.66,-1.0),
(0.66,-1.0,1.0,-5.0/6.0),
(1.0,-5.0/6.0,0.5,-0.5),
(0.5,0,0.5,-1.0/6.0),
(0.5,-0.33,0.5,-0.5),
]]
def lbkt(self):
x,y,x2,y2 = self.ns(0.5)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,0,0,-0.33),
(0,-0.33,0,-0.66),
(0,-0.66,1.0,-1.0),
]]
def rbkt(self):
x,y,x2,y2 = self.ns(0.5)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,1.0,-0.33),
(1.0,-0.33,1.0,-0.66),
(1.0,-0.66,0,-1.0),
]]
def star(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,1.0,-0.66),
(1.0,-0.33,0,-0.66),
(0,-0.5,1.0,-0.5),
]]
def dr(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,1.0,-1.0),
]]
def plus(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0.5,-0.33,0.5,-0.66),
(0,-0.5,1.0,-0.5),
]]
def minus(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.5,1.0,-0.5),
]]
def equal(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.33,1.0,-0.33),
(0,-0.66,1.0,-0.66),
]]
def smollest(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,-0.5,1.0,-0.33),
(0,-0.5,1.0,-0.66),
]]
def bigest(self):
x,y,x2,y2 = self.ns(0.8)
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(1.0,-0.5,0,-0.33),
(1.0,-0.5,0,-0.66),
]]
def num(self):
x,y,x2,y2 = self.ns()
return [ [x+x2*ax, y+y2*ay, x+x2*bx, y+y2*by] for (ax,ay,bx,by) in [
(0,0,0,-1.0),
(1.0,0,1.0,-1.0),
(0,-1.0,1.0,0),
(7.0/6.0,-1.0,1.33,-1.0),
(1.33,-1.0,1.33,-0.75),
(1.33,-0.75,7.0/6.0,-0.75),
(7.0/6.0,-0.75,7.0/6.0,-1.0),
(1.33,-0.5,7.0/6.0,-0.5),
]]
| 27.984714 | 132 | 0.342018 | 15,001 | 78,721 | 1.790614 | 0.021065 | 0.102007 | 0.085552 | 0.051971 | 0.888649 | 0.836938 | 0.806001 | 0.765087 | 0.755668 | 0.738357 | 0 | 0.18821 | 0.406321 | 78,721 | 2,812 | 133 | 27.994666 | 0.38637 | 0.011725 | 0 | 0.798845 | 0 | 0 | 0.004709 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.096246 | false | 0 | 0.000481 | 0 | 0.192974 | 0 | 0 | 0 | 1 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 9 |
87a47ac74561dc50468225283ac64ebcb6a54698 | 32,272 | py | Python | mailslurp_client/api/domain_controller_api.py | mailslurp/mailslurp-client-python | a1e9fdc6eb06e192909fd57a64813beb32419594 | [
"MIT"
] | 6 | 2020-04-30T07:47:42.000Z | 2022-03-24T20:58:58.000Z | mailslurp_client/api/domain_controller_api.py | mailslurp/mailslurp-client-python | a1e9fdc6eb06e192909fd57a64813beb32419594 | [
"MIT"
] | 1 | 2020-09-20T19:58:21.000Z | 2020-11-29T16:49:19.000Z | mailslurp_client/api/domain_controller_api.py | mailslurp/mailslurp-client-python | a1e9fdc6eb06e192909fd57a64813beb32419594 | [
"MIT"
] | 1 | 2019-08-09T14:55:50.000Z | 2019-08-09T14:55:50.000Z | # coding: utf-8
"""
MailSlurp API
MailSlurp is an API for sending and receiving emails from dynamically allocated email addresses. It's designed for developers and QA teams to test applications, process inbound emails, send templated notifications, attachments, and more. ## Resources - [Homepage](https://www.mailslurp.com) - Get an [API KEY](https://app.mailslurp.com/sign-up/) - Generated [SDK Clients](https://www.mailslurp.com/docs/) - [Examples](https://github.com/mailslurp/examples) repository # noqa: E501
The version of the OpenAPI document: 6.5.2
Generated by: https://openapi-generator.tech
"""
from __future__ import absolute_import
import re # noqa: F401
# python 2 and python 3 compatibility library
import six
from mailslurp_client.api_client import ApiClient
from mailslurp_client.exceptions import ( # noqa: F401
ApiTypeError,
ApiValueError
)
class DomainControllerApi(object):
"""NOTE: This class is auto generated by OpenAPI Generator
Ref: https://openapi-generator.tech
Do not edit the class manually.
"""
def __init__(self, api_client=None):
if api_client is None:
api_client = ApiClient()
self.api_client = api_client
def add_domain_wildcard_catch_all(self, id, **kwargs): # noqa: E501
"""Add catch all wild card inbox to domain # noqa: E501
Add a catch all inbox to a domain so that any emails sent to it that cannot be matched will be sent to the catch all inbox generated # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.add_domain_wildcard_catch_all(id, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param str id: id (required)
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: DomainDto
If the method is called asynchronously,
returns the request thread.
"""
kwargs['_return_http_data_only'] = True
return self.add_domain_wildcard_catch_all_with_http_info(id, **kwargs) # noqa: E501
def add_domain_wildcard_catch_all_with_http_info(self, id, **kwargs): # noqa: E501
"""Add catch all wild card inbox to domain # noqa: E501
Add a catch all inbox to a domain so that any emails sent to it that cannot be matched will be sent to the catch all inbox generated # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.add_domain_wildcard_catch_all_with_http_info(id, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param str id: id (required)
:param _return_http_data_only: response data without head status code
and headers
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: tuple(DomainDto, status_code(int), headers(HTTPHeaderDict))
If the method is called asynchronously,
returns the request thread.
"""
local_var_params = locals()
all_params = [
'id'
]
all_params.extend(
[
'async_req',
'_return_http_data_only',
'_preload_content',
'_request_timeout'
]
)
for key, val in six.iteritems(local_var_params['kwargs']):
if key not in all_params:
raise ApiTypeError(
"Got an unexpected keyword argument '%s'"
" to method add_domain_wildcard_catch_all" % key
)
local_var_params[key] = val
del local_var_params['kwargs']
# verify the required parameter 'id' is set
if self.api_client.client_side_validation and ('id' not in local_var_params or # noqa: E501
local_var_params['id'] is None): # noqa: E501
raise ApiValueError("Missing the required parameter `id` when calling `add_domain_wildcard_catch_all`") # noqa: E501
collection_formats = {}
path_params = {}
if 'id' in local_var_params:
path_params['id'] = local_var_params['id'] # noqa: E501
query_params = []
header_params = {}
form_params = []
local_var_files = {}
body_params = None
# HTTP header `Accept`
header_params['Accept'] = self.api_client.select_header_accept(
['application/json']) # noqa: E501
# Authentication setting
auth_settings = ['API_KEY'] # noqa: E501
return self.api_client.call_api(
'/domains/{id}/wildcard', 'POST',
path_params,
query_params,
header_params,
body=body_params,
post_params=form_params,
files=local_var_files,
response_type='DomainDto', # noqa: E501
auth_settings=auth_settings,
async_req=local_var_params.get('async_req'),
_return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501
_preload_content=local_var_params.get('_preload_content', True),
_request_timeout=local_var_params.get('_request_timeout'),
collection_formats=collection_formats)
def create_domain(self, domain_options, **kwargs): # noqa: E501
"""Create Domain # noqa: E501
Link a domain that you own with MailSlurp so you can create email addresses using it. Endpoint returns DNS records used for validation. You must add these verification records to your host provider's DNS setup to verify the domain. # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.create_domain(domain_options, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param CreateDomainOptions domain_options: domainOptions (required)
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: DomainDto
If the method is called asynchronously,
returns the request thread.
"""
kwargs['_return_http_data_only'] = True
return self.create_domain_with_http_info(domain_options, **kwargs) # noqa: E501
def create_domain_with_http_info(self, domain_options, **kwargs): # noqa: E501
"""Create Domain # noqa: E501
Link a domain that you own with MailSlurp so you can create email addresses using it. Endpoint returns DNS records used for validation. You must add these verification records to your host provider's DNS setup to verify the domain. # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.create_domain_with_http_info(domain_options, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param CreateDomainOptions domain_options: domainOptions (required)
:param _return_http_data_only: response data without head status code
and headers
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: tuple(DomainDto, status_code(int), headers(HTTPHeaderDict))
If the method is called asynchronously,
returns the request thread.
"""
local_var_params = locals()
all_params = [
'domain_options'
]
all_params.extend(
[
'async_req',
'_return_http_data_only',
'_preload_content',
'_request_timeout'
]
)
for key, val in six.iteritems(local_var_params['kwargs']):
if key not in all_params:
raise ApiTypeError(
"Got an unexpected keyword argument '%s'"
" to method create_domain" % key
)
local_var_params[key] = val
del local_var_params['kwargs']
# verify the required parameter 'domain_options' is set
if self.api_client.client_side_validation and ('domain_options' not in local_var_params or # noqa: E501
local_var_params['domain_options'] is None): # noqa: E501
raise ApiValueError("Missing the required parameter `domain_options` when calling `create_domain`") # noqa: E501
collection_formats = {}
path_params = {}
query_params = []
header_params = {}
form_params = []
local_var_files = {}
body_params = None
if 'domain_options' in local_var_params:
body_params = local_var_params['domain_options']
# HTTP header `Accept`
header_params['Accept'] = self.api_client.select_header_accept(
['application/json']) # noqa: E501
# HTTP header `Content-Type`
header_params['Content-Type'] = self.api_client.select_header_content_type( # noqa: E501
['application/json']) # noqa: E501
# Authentication setting
auth_settings = ['API_KEY'] # noqa: E501
return self.api_client.call_api(
'/domains', 'POST',
path_params,
query_params,
header_params,
body=body_params,
post_params=form_params,
files=local_var_files,
response_type='DomainDto', # noqa: E501
auth_settings=auth_settings,
async_req=local_var_params.get('async_req'),
_return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501
_preload_content=local_var_params.get('_preload_content', True),
_request_timeout=local_var_params.get('_request_timeout'),
collection_formats=collection_formats)
def delete_domain(self, id, **kwargs): # noqa: E501
"""Delete a domain # noqa: E501
Delete a domain. This will disable any existing inboxes that use this domain. # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.delete_domain(id, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param str id: id (required)
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: list[str]
If the method is called asynchronously,
returns the request thread.
"""
kwargs['_return_http_data_only'] = True
return self.delete_domain_with_http_info(id, **kwargs) # noqa: E501
def delete_domain_with_http_info(self, id, **kwargs): # noqa: E501
"""Delete a domain # noqa: E501
Delete a domain. This will disable any existing inboxes that use this domain. # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.delete_domain_with_http_info(id, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param str id: id (required)
:param _return_http_data_only: response data without head status code
and headers
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: tuple(list[str], status_code(int), headers(HTTPHeaderDict))
If the method is called asynchronously,
returns the request thread.
"""
local_var_params = locals()
all_params = [
'id'
]
all_params.extend(
[
'async_req',
'_return_http_data_only',
'_preload_content',
'_request_timeout'
]
)
for key, val in six.iteritems(local_var_params['kwargs']):
if key not in all_params:
raise ApiTypeError(
"Got an unexpected keyword argument '%s'"
" to method delete_domain" % key
)
local_var_params[key] = val
del local_var_params['kwargs']
# verify the required parameter 'id' is set
if self.api_client.client_side_validation and ('id' not in local_var_params or # noqa: E501
local_var_params['id'] is None): # noqa: E501
raise ApiValueError("Missing the required parameter `id` when calling `delete_domain`") # noqa: E501
collection_formats = {}
path_params = {}
if 'id' in local_var_params:
path_params['id'] = local_var_params['id'] # noqa: E501
query_params = []
header_params = {}
form_params = []
local_var_files = {}
body_params = None
# HTTP header `Accept`
header_params['Accept'] = self.api_client.select_header_accept(
['application/json']) # noqa: E501
# Authentication setting
auth_settings = ['API_KEY'] # noqa: E501
return self.api_client.call_api(
'/domains/{id}', 'DELETE',
path_params,
query_params,
header_params,
body=body_params,
post_params=form_params,
files=local_var_files,
response_type='list[str]', # noqa: E501
auth_settings=auth_settings,
async_req=local_var_params.get('async_req'),
_return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501
_preload_content=local_var_params.get('_preload_content', True),
_request_timeout=local_var_params.get('_request_timeout'),
collection_formats=collection_formats)
def get_domain(self, id, **kwargs): # noqa: E501
"""Get a domain # noqa: E501
Returns domain verification status and tokens for a given domain # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.get_domain(id, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param str id: id (required)
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: DomainDto
If the method is called asynchronously,
returns the request thread.
"""
kwargs['_return_http_data_only'] = True
return self.get_domain_with_http_info(id, **kwargs) # noqa: E501
def get_domain_with_http_info(self, id, **kwargs): # noqa: E501
"""Get a domain # noqa: E501
Returns domain verification status and tokens for a given domain # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.get_domain_with_http_info(id, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param str id: id (required)
:param _return_http_data_only: response data without head status code
and headers
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: tuple(DomainDto, status_code(int), headers(HTTPHeaderDict))
If the method is called asynchronously,
returns the request thread.
"""
local_var_params = locals()
all_params = [
'id'
]
all_params.extend(
[
'async_req',
'_return_http_data_only',
'_preload_content',
'_request_timeout'
]
)
for key, val in six.iteritems(local_var_params['kwargs']):
if key not in all_params:
raise ApiTypeError(
"Got an unexpected keyword argument '%s'"
" to method get_domain" % key
)
local_var_params[key] = val
del local_var_params['kwargs']
# verify the required parameter 'id' is set
if self.api_client.client_side_validation and ('id' not in local_var_params or # noqa: E501
local_var_params['id'] is None): # noqa: E501
raise ApiValueError("Missing the required parameter `id` when calling `get_domain`") # noqa: E501
collection_formats = {}
path_params = {}
if 'id' in local_var_params:
path_params['id'] = local_var_params['id'] # noqa: E501
query_params = []
header_params = {}
form_params = []
local_var_files = {}
body_params = None
# HTTP header `Accept`
header_params['Accept'] = self.api_client.select_header_accept(
['application/json']) # noqa: E501
# Authentication setting
auth_settings = ['API_KEY'] # noqa: E501
return self.api_client.call_api(
'/domains/{id}', 'GET',
path_params,
query_params,
header_params,
body=body_params,
post_params=form_params,
files=local_var_files,
response_type='DomainDto', # noqa: E501
auth_settings=auth_settings,
async_req=local_var_params.get('async_req'),
_return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501
_preload_content=local_var_params.get('_preload_content', True),
_request_timeout=local_var_params.get('_request_timeout'),
collection_formats=collection_formats)
def get_domains(self, **kwargs): # noqa: E501
"""Get domains # noqa: E501
List all custom domains you have created # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.get_domains(async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: list[DomainPreview]
If the method is called asynchronously,
returns the request thread.
"""
kwargs['_return_http_data_only'] = True
return self.get_domains_with_http_info(**kwargs) # noqa: E501
def get_domains_with_http_info(self, **kwargs): # noqa: E501
"""Get domains # noqa: E501
List all custom domains you have created # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.get_domains_with_http_info(async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param _return_http_data_only: response data without head status code
and headers
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: tuple(list[DomainPreview], status_code(int), headers(HTTPHeaderDict))
If the method is called asynchronously,
returns the request thread.
"""
local_var_params = locals()
all_params = [
]
all_params.extend(
[
'async_req',
'_return_http_data_only',
'_preload_content',
'_request_timeout'
]
)
for key, val in six.iteritems(local_var_params['kwargs']):
if key not in all_params:
raise ApiTypeError(
"Got an unexpected keyword argument '%s'"
" to method get_domains" % key
)
local_var_params[key] = val
del local_var_params['kwargs']
collection_formats = {}
path_params = {}
query_params = []
header_params = {}
form_params = []
local_var_files = {}
body_params = None
# HTTP header `Accept`
header_params['Accept'] = self.api_client.select_header_accept(
['application/json']) # noqa: E501
# Authentication setting
auth_settings = ['API_KEY'] # noqa: E501
return self.api_client.call_api(
'/domains', 'GET',
path_params,
query_params,
header_params,
body=body_params,
post_params=form_params,
files=local_var_files,
response_type='list[DomainPreview]', # noqa: E501
auth_settings=auth_settings,
async_req=local_var_params.get('async_req'),
_return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501
_preload_content=local_var_params.get('_preload_content', True),
_request_timeout=local_var_params.get('_request_timeout'),
collection_formats=collection_formats)
def update_domain(self, id, update_domain_dto, **kwargs): # noqa: E501
"""Update a domain # noqa: E501
Update values on a domain. Note you cannot change the domain name as it is immutable. Recreate the domain if you need to alter this. # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.update_domain(id, update_domain_dto, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param str id: id (required)
:param UpdateDomainOptions update_domain_dto: updateDomainDto (required)
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: DomainDto
If the method is called asynchronously,
returns the request thread.
"""
kwargs['_return_http_data_only'] = True
return self.update_domain_with_http_info(id, update_domain_dto, **kwargs) # noqa: E501
def update_domain_with_http_info(self, id, update_domain_dto, **kwargs): # noqa: E501
"""Update a domain # noqa: E501
Update values on a domain. Note you cannot change the domain name as it is immutable. Recreate the domain if you need to alter this. # noqa: E501
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please pass async_req=True
>>> thread = api.update_domain_with_http_info(id, update_domain_dto, async_req=True)
>>> result = thread.get()
:param async_req bool: execute request asynchronously
:param str id: id (required)
:param UpdateDomainOptions update_domain_dto: updateDomainDto (required)
:param _return_http_data_only: response data without head status code
and headers
:param _preload_content: if False, the urllib3.HTTPResponse object will
be returned without reading/decoding response
data. Default is True.
:param _request_timeout: timeout setting for this request. If one
number provided, it will be total request
timeout. It can also be a pair (tuple) of
(connection, read) timeouts.
:return: tuple(DomainDto, status_code(int), headers(HTTPHeaderDict))
If the method is called asynchronously,
returns the request thread.
"""
local_var_params = locals()
all_params = [
'id',
'update_domain_dto'
]
all_params.extend(
[
'async_req',
'_return_http_data_only',
'_preload_content',
'_request_timeout'
]
)
for key, val in six.iteritems(local_var_params['kwargs']):
if key not in all_params:
raise ApiTypeError(
"Got an unexpected keyword argument '%s'"
" to method update_domain" % key
)
local_var_params[key] = val
del local_var_params['kwargs']
# verify the required parameter 'id' is set
if self.api_client.client_side_validation and ('id' not in local_var_params or # noqa: E501
local_var_params['id'] is None): # noqa: E501
raise ApiValueError("Missing the required parameter `id` when calling `update_domain`") # noqa: E501
# verify the required parameter 'update_domain_dto' is set
if self.api_client.client_side_validation and ('update_domain_dto' not in local_var_params or # noqa: E501
local_var_params['update_domain_dto'] is None): # noqa: E501
raise ApiValueError("Missing the required parameter `update_domain_dto` when calling `update_domain`") # noqa: E501
collection_formats = {}
path_params = {}
if 'id' in local_var_params:
path_params['id'] = local_var_params['id'] # noqa: E501
query_params = []
header_params = {}
form_params = []
local_var_files = {}
body_params = None
if 'update_domain_dto' in local_var_params:
body_params = local_var_params['update_domain_dto']
# HTTP header `Accept`
header_params['Accept'] = self.api_client.select_header_accept(
['application/json']) # noqa: E501
# HTTP header `Content-Type`
header_params['Content-Type'] = self.api_client.select_header_content_type( # noqa: E501
['application/json']) # noqa: E501
# Authentication setting
auth_settings = ['API_KEY'] # noqa: E501
return self.api_client.call_api(
'/domains/{id}', 'PUT',
path_params,
query_params,
header_params,
body=body_params,
post_params=form_params,
files=local_var_files,
response_type='DomainDto', # noqa: E501
auth_settings=auth_settings,
async_req=local_var_params.get('async_req'),
_return_http_data_only=local_var_params.get('_return_http_data_only'), # noqa: E501
_preload_content=local_var_params.get('_preload_content', True),
_request_timeout=local_var_params.get('_request_timeout'),
collection_formats=collection_formats)
| 44.208219 | 487 | 0.589768 | 3,621 | 32,272 | 5.022646 | 0.071251 | 0.040908 | 0.055424 | 0.029692 | 0.938143 | 0.925771 | 0.915819 | 0.911585 | 0.908616 | 0.89311 | 0 | 0.014193 | 0.338498 | 32,272 | 729 | 488 | 44.268861 | 0.837737 | 0.477132 | 0 | 0.730205 | 0 | 0 | 0.15965 | 0.032199 | 0 | 0 | 0 | 0 | 0 | 1 | 0.038123 | false | 0 | 0.014663 | 0 | 0.090909 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
87eba730d3fc393d69e8c1f01825615681524d41 | 2,563 | py | Python | payment/filter.py | AhmedElmawary/erp | c998787c62194e26e10e3cbc61e35935e901e56d | [
"MIT"
] | null | null | null | payment/filter.py | AhmedElmawary/erp | c998787c62194e26e10e3cbc61e35935e901e56d | [
"MIT"
] | null | null | null | payment/filter.py | AhmedElmawary/erp | c998787c62194e26e10e3cbc61e35935e901e56d | [
"MIT"
] | null | null | null | from django.contrib import admin
from django.utils.translation import gettext_lazy
from payment.models import PaymentTransactionType
class ClientTransactionType(admin.SimpleListFilter):
# Human-readable title which will be displayed in the
# right admin sidebar just above the filter options.
title = gettext_lazy('Transaction type')
# Parameter for the filter that will be used in the URL query.
# parameter_name = 'type_tranasction__id__exact'
parameter_name = 'transactions_type'
def lookups(self, request, model_admin):
"""
Returns a list of tuples. The first element in each
tuple is the coded value for the option that will
appear in the URL query. The second element is the
human-readable name for the option that will appear
in the right sidebar.
"""
options_for_client = []
for row in PaymentTransactionType.objects.filter(for_client=True):
options_for_client.append((row.id, str(row)))
return options_for_client
def queryset(self, request, queryset):
"""
Returns the filtered queryset based on the value
provided in the query string and retrievable via
`self.value()`.
"""
if self.value():
return queryset.filter(type_tranasction=self.value())
class SupplierTransactionType(admin.SimpleListFilter):
# Human-readable title which will be displayed in the
# right admin sidebar just above the filter options.
title = gettext_lazy('Transaction type')
# Parameter for the filter that will be used in the URL query.
# parameter_name = 'type_tranasction__id__exact'
parameter_name = 'transactions_type'
def lookups(self, request, model_admin):
"""
Returns a list of tuples. The first element in each
tuple is the coded value for the option that will
appear in the URL query. The second element is the
human-readable name for the option that will appear
in the right sidebar.
"""
options_for_supplier = []
for row in PaymentTransactionType.objects.filter(for_supplier=True):
options_for_supplier.append((row.id, str(row)))
return options_for_supplier
def queryset(self, request, queryset):
"""
Returns the filtered queryset based on the value
provided in the query string and retrievable via
`self.value()`.
"""
if self.value():
return queryset.filter(type_tranasction=self.value())
| 37.691176 | 76 | 0.680062 | 325 | 2,563 | 5.249231 | 0.252308 | 0.029308 | 0.023447 | 0.030481 | 0.851114 | 0.851114 | 0.851114 | 0.797186 | 0.758499 | 0.758499 | 0 | 0 | 0.253609 | 2,563 | 67 | 77 | 38.253731 | 0.891793 | 0.430745 | 0 | 0.48 | 0 | 0 | 0.052257 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.16 | false | 0 | 0.12 | 0 | 0.68 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 7 |
3576c087b9c0cfbb0f42b8bc708af9a24746ec4d | 14,707 | py | Python | ScriptArchive/LSTM_multi_detect_FranklinBasin.py | AmberSJones/PyHydroQC | 9f8992672ce3163eb048964e85680c526b4fd3f3 | [
"BSD-3-Clause"
] | 6 | 2021-04-22T15:12:55.000Z | 2022-02-01T00:47:15.000Z | ScriptArchive/LSTM_multi_detect_FranklinBasin.py | AmberSJones/PyHydroQC | 9f8992672ce3163eb048964e85680c526b4fd3f3 | [
"BSD-3-Clause"
] | 6 | 2020-12-22T14:12:30.000Z | 2021-04-02T22:53:47.000Z | ScriptArchive/LSTM_multi_detect_FranklinBasin.py | AmberSJones/PyHydroQC | 9f8992672ce3163eb048964e85680c526b4fd3f3 | [
"BSD-3-Clause"
] | 6 | 2021-01-06T08:09:59.000Z | 2022-02-01T00:46:22.000Z | ################################
# LSTM DEVELOP AND DIAGNOSTIC #
################################
# This code takes raw data and corrected data, applies an LSTM model, and identifies anomalies.
import rules_detect
import anomaly_utilities
import modeling_utilities
import numpy as np
import tensorflow as tf
import pandas as pd
import seaborn as sns
from matplotlib.pylab import rcParams
import matplotlib.pyplot as plt
import plotly.io as pio
pio.renderers.default = "browser"
pd.options.mode.chained_assignment = None
sns.set(style='whitegrid', palette='muted')
rcParams['figure.figsize'] = 14, 8
np.random.seed(1)
print('Tensorflow version:', tf.__version__)
print("LSTM exploration script begin.")
##################################################
# LSTM Multivariate Retrieve and Preprocess Data #
##################################################
# DEFINE SITE and VARIABLE #
#########################################
# site = "BlackSmithFork"
site = "FranklinBasin"
# site = "MainStreet"
# site = "Mendon"
# site = "TonyGrove"
# site = "WaterLab"
sensor = ['temp', 'cond', 'ph', 'do']
year = [2014, 2015, 2016, 2017, 2018, 2019]
# GET DATA #
#########################################
df_full, sensor_array = anomaly_utilities.get_data(site, sensor, year, path="/LRO_data/")
# RULES BASED DETECTION #
#########################################
maximum = [13, 380, 9.2, 13]
minimum = [-2, 120, 7.5, 8]
length = 6
size = []
for i in range(0, len(sensor_array)):
sensor_array[sensor[i]] = rules_detect.range_check(sensor_array[sensor[i]], maximum[i], minimum[i])
sensor_array[sensor[i]] = rules_detect.persistence(sensor_array[sensor[i]], length)
s = rules_detect.group_size(sensor_array[sensor[i]])
size.append(s)
sensor_array[sensor[i]] = rules_detect.interpolate(sensor_array[sensor[i]])
# Create new data frame with raw and corrected data for variables of interest
df_observed = pd.DataFrame(index=df_full.index)
df_observed['temp_obs'] = sensor_array['temp']['observed']
df_observed['cond_obs'] = sensor_array['cond']['observed']
df_observed['ph_obs'] = sensor_array['ph']['observed']
df_observed['do_obs'] = sensor_array['do']['observed']
df_raw = pd.DataFrame(index=df_full.index)
df_raw['temp'] = df_full['temp']
df_raw['cond'] = df_full['cond']
df_raw['ph'] = df_full['ph']
df_raw['do'] = df_full['do']
df_anomaly = pd.DataFrame(index=df_full.index)
df_anomaly['temp_anom'] = sensor_array['temp']['anomaly']
df_anomaly['cond_anom'] = sensor_array['cond']['anomaly']
df_anomaly['ph_anom'] = sensor_array['ph']['anomaly']
df_anomaly['do_anom'] = sensor_array['do']['anomaly']
print(df_observed.shape)
print(df_raw.shape)
print(df_anomaly.shape)
#########################################
# LSTM Multivariate Vanilla Model #
#########################################
# MODEL CREATION #
#########################################
# scales data, reshapes data, builds and trains model, evaluates model results
time_steps = 10
samples = 10000
cells = 128
dropout = 0.2
patience = 6
lstm_multivar = modeling_utilities.lstm_multivar(df_observed, df_anomaly, df_raw, time_steps, samples, cells, dropout, patience)
# Plot Metrics and Evaluate the Model
# plot training loss and validation loss with matplotlib and pyplot
plt.plot(lstm_multivar.history.history['loss'], label='Training Loss')
plt.plot(lstm_multivar.history.history['val_loss'], label='Validation Loss')
plt.legend()
plt.show()
# DETERMINE THRESHOLD AND DETECT ANOMALIES #
#########################################
residuals = pd.DataFrame(lstm_multivar.test_residuals)
predictions = pd.DataFrame(lstm_multivar.predictions)
residuals.index = df_observed[time_steps:].index
predictions.index = df_observed[time_steps:].index
window_sz = [40, 40, 40, 40]
alpha = [0.0001, 0.0001, 0.0001, 0.001]
min_range = [0.25, 5, 0.01, 0.15]
threshold = []
for i in range(0, lstm_multivar.test_residuals.shape[1]):
threshold_df = anomaly_utilities.set_dynamic_threshold(residuals.iloc[:, i], window_sz[i], alpha[i], min_range[i])
threshold_df.index = residuals.index
threshold.append(threshold_df)
plt.figure()
# plt.plot(df['raw'], 'b', label='original data')
plt.plot(residuals.iloc[:, i], 'b', label='residuals')
plt.plot(threshold[i]['low'], 'c', label='thresh_low')
plt.plot(threshold[i]['high'], 'm', mfc='none', label='thresh_high')
plt.legend()
plt.ylabel(sensor[i])
plt.show()
observed = df_observed[time_steps:]
detections_array = []
for i in range(0, observed.shape[1]):
detections_df = anomaly_utilities.detect_anomalies(observed.iloc[:, i], lstm_multivar.predictions.iloc[:, i], lstm_multivar.test_residuals.iloc[:, i], threshold[i], summary=True)
detections_array.append(detections_df)
# Use events function to widen and number anomalous events
df_array = []
for i in range(0, len(detections_array)):
all_data = []
all_data = sensor_array[sensor[i]].iloc[time_steps:]
all_data['labeled_event'] = anomaly_utilities.anomaly_events(all_data['labeled_anomaly'])
all_data['detected_anomaly'] = detections_array[i]['anomaly']
all_data['detected_event'] = anomaly_utilities.anomaly_events(all_data['detected_anomaly'])
df_array.append(all_data)
# DETERMINE METRICS #
#########################################
anomaly_utilities.compare_events(df_array[0])
temp_metrics = anomaly_utilities.metrics(df_array[0])
anomaly_utilities.compare_events(df_array[1])
cond_metrics = anomaly_utilities.metrics(df_array[1])
anomaly_utilities.compare_events(df_array[2])
ph_metrics = anomaly_utilities.metrics(df_array[2])
anomaly_utilities.compare_events(df_array[3])
do_metrics = anomaly_utilities.metrics(df_array[3])
# OUTPUT RESULTS #
#########################################
print('\n\n\nScript report:\n')
print('Sensor: temp')
print('Year: ' + str(year))
# print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold))
print('PPV = %f' % temp_metrics.prc)
print('NPV = %f' % temp_metrics.npv)
print('Acc = %f' % temp_metrics.acc)
print('TP = %i' % temp_metrics.true_positives)
print('TN = %i' % temp_metrics.true_negatives)
print('FP = %i' % temp_metrics.false_positives)
print('FN = %i' % temp_metrics.false_negatives)
print('F1 = %f' % temp_metrics.f1)
print('F2 = %f' % temp_metrics.f2)
print('\n\n\nScript report:\n')
print('Sensor: cond')
print('Year: ' + str(year))
# print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold))
print('PPV = %f' % cond_metrics.prc)
print('NPV = %f' % cond_metrics.npv)
print('Acc = %f' % cond_metrics.acc)
print('TP = %i' % cond_metrics.true_positives)
print('TN = %i' % cond_metrics.true_negatives)
print('FP = %i' % cond_metrics.false_positives)
print('FN = %i' % cond_metrics.false_negatives)
print('F1 = %f' % cond_metrics.f1)
print('F2 = %f' % cond_metrics.f2)
print('\n\n\nScript report:\n')
print('Sensor: ph')
print('Year: ' + str(year))
# print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold))
print('PPV = %f' % ph_metrics.prc)
print('NPV = %f' % ph_metrics.npv)
print('Acc = %f' % ph_metrics.acc)
print('TP = %i' % ph_metrics.true_positives)
print('TN = %i' % ph_metrics.true_negatives)
print('FP = %i' % ph_metrics.false_positives)
print('FN = %i' % ph_metrics.false_negatives)
print('F1 = %f' % ph_metrics.f1)
print('F2 = %f' % ph_metrics.f2)
print('\n\n\nScript report:\n')
print('Sensor: do')
print('Year: ' + str(year))
# print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold))
print('PPV = %f' % do_metrics.prc)
print('NPV = %f' % do_metrics.npv)
print('Acc = %f' % do_metrics.acc)
print('TP = %i' % do_metrics.true_positives)
print('TN = %i' % do_metrics.true_negatives)
print('FP = %i' % do_metrics.false_positives)
print('FN = %i' % do_metrics.false_negatives)
print('F1 = %f' % do_metrics.f1)
print('F2 = %f' % do_metrics.f2)
# GENERATE PLOTS #
#########################################
for i in range(0, len(sensor)):
plt.figure()
plt.plot(df_raw[df_raw.columns[i]], 'b', label='original data')
#plt.plot(df_observed[df_observed.columns[i]], 'm', label='corrected data' )
plt.plot(detections_array[i]['prediction'], 'c', label='predicted values')
plt.plot(sensor_array[sensor[i]]['raw'][sensor_array[sensor[i]]['labeled_anomaly']], 'mo', mfc='none', label='technician labeled anomalies')
plt.plot(detections_array[i]['prediction'][detections_array[i]['anomaly']], 'r+', label='machine detected anomalies')
plt.legend()
plt.ylabel(sensor[i])
plt.show()
#########################################
# LSTM Multivariate Bidirectional Model #
#########################################
# Model creation #
#########################################
# scales data, reshapes data, builds and trains model, evaluates model results
time_steps = 10
samples = 10000
cells = 128
dropout = 0.2
patience = 6
lstm_multivar_bidir = modeling_utilities.lstm_multivar_bidir(df_observed, df_anomaly, df_raw, time_steps, samples, cells, dropout, patience)
# Plot Metrics and Evaluate the Model
# plot training loss and validation loss with matplotlib and pyplot
plt.plot(lstm_multivar_bidir.history.history['loss'], label='Training Loss')
plt.plot(lstm_multivar_bidir.history.history['val_loss'], label='Validation Loss')
plt.legend()
plt.show()
# DETERMINE THRESHOLD AND DETECT ANOMALIES #
#########################################
residuals = pd.DataFrame(lstm_multivar_bidir.test_residuals)
residuals.index = df_observed[time_steps:-time_steps].index
window_sz = [40, 40, 40, 40]
alpha = [0.0001, 0.0001, 0.0001, 0.001]
min_range = [0.25, 5, 0.01, 0.15]
threshold = []
for i in range(0, lstm_multivar_bidir.test_residuals.shape[1]):
threshold_df = anomaly_utilities.set_dynamic_threshold(residuals.iloc[:, i], window_sz[i], alpha[i], min_range[i])
threshold_df.index = residuals.index
threshold.append(threshold_df)
plt.figure()
# plt.plot(df['raw'], 'b', label='original data')
plt.plot(residuals.iloc[:, i], 'b', label='residuals')
plt.plot(threshold[i]['low'], 'c', label='thresh_low')
plt.plot(threshold[i]['high'], 'm', mfc='none', label='thresh_high')
plt.legend()
plt.ylabel(sensor[i])
plt.show()
observed = df_observed[time_steps:-time_steps]
detections_array = []
for i in range(0, observed.shape[1]):
detections_df = anomaly_utilities.detect_anomalies(observed.iloc[:, i], lstm_multivar_bidir.predictions.iloc[:, i], lstm_multivar_bidir.test_residuals.iloc[:, i], threshold[i], summary=True)
detections_array.append(detections_df)
# Use events function to widen and number anomalous events
df_array = []
for i in range(0, len(detections_array)):
all_data = []
all_data = sensor_array[sensor[i]].iloc[time_steps:]
all_data['labeled_event'] = anomaly_utilities.anomaly_events(all_data['labeled_anomaly'])
all_data['detected_anomaly'] = detections_array[i]['anomaly']
all_data['detected_event'] = anomaly_utilities.anomaly_events(all_data['detected_anomaly'])
df_array.append(all_data)
# DETERMINE METRICS #
#########################################
anomaly_utilities.compare_events(df_array[0])
temp_metrics = anomaly_utilities.metrics(df_array[0])
anomaly_utilities.compare_events(df_array[1])
cond_metrics = anomaly_utilities.metrics(df_array[1])
anomaly_utilities.compare_events(df_array[2])
ph_metrics = anomaly_utilities.metrics(df_array[2])
anomaly_utilities.compare_events(df_array[3])
do_metrics = anomaly_utilities.metrics(df_array[3])
# OUTPUT RESULTS #
#########################################
print('\n\n\nScript report:\n')
print('Sensor: temp')
print('Year: ' + str(year))
# print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold))
print('PPV = %f' % temp_metrics.prc)
print('NPV = %f' % temp_metrics.npv)
print('Acc = %f' % temp_metrics.acc)
print('TP = %i' % temp_metrics.true_positives)
print('TN = %i' % temp_metrics.true_negatives)
print('FP = %i' % temp_metrics.false_positives)
print('FN = %i' % temp_metrics.false_negatives)
print('F1 = %f' % temp_metrics.f1)
print('F2 = %f' % temp_metrics.f2)
print('\n\n\nScript report:\n')
print('Sensor: cond')
print('Year: ' + str(year))
# print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold))
print('PPV = %f' % cond_metrics.prc)
print('NPV = %f' % cond_metrics.npv)
print('Acc = %f' % cond_metrics.acc)
print('TP = %i' % cond_metrics.true_positives)
print('TN = %i' % cond_metrics.true_negatives)
print('FP = %i' % cond_metrics.false_positives)
print('FN = %i' % cond_metrics.false_negatives)
print('F1 = %f' % cond_metrics.f1)
print('F2 = %f' % cond_metrics.f2)
print('\n\n\nScript report:\n')
print('Sensor: ph')
print('Year: ' + str(year))
# print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold))
print('PPV = %f' % ph_metrics.prc)
print('NPV = %f' % ph_metrics.npv)
print('Acc = %f' % ph_metrics.acc)
print('TP = %i' % ph_metrics.true_positives)
print('TN = %i' % ph_metrics.true_negatives)
print('FP = %i' % ph_metrics.false_positives)
print('FN = %i' % ph_metrics.false_negatives)
print('F1 = %f' % ph_metrics.f1)
print('F2 = %f' % ph_metrics.f2)
print('\n\n\nScript report:\n')
print('Sensor: do')
print('Year: ' + str(year))
# print('Parameters: LSTM, sequence length: %i, training samples: %i, Threshold = %f' %(time_steps, samples, threshold))
print('PPV = %f' % do_metrics.prc)
print('NPV = %f' % do_metrics.npv)
print('Acc = %f' % do_metrics.acc)
print('TP = %i' % do_metrics.true_positives)
print('TN = %i' % do_metrics.true_negatives)
print('FP = %i' % do_metrics.false_positives)
print('FN = %i' % do_metrics.false_negatives)
print('F1 = %f' % do_metrics.f1)
print('F2 = %f' % do_metrics.f2)
# GENERATE PLOTS #
#########################################
for i in range(0, len(sensor)):
plt.figure()
plt.plot(df_raw[df_raw.columns[i]], 'b', label='original data')
# plt.plot(df_observed[df_observed.columns[i]], 'm', label='corrected data' )
plt.plot(detections_array[i]['prediction'], 'c', label='predicted values')
plt.plot(sensor_array[sensor[i]]['raw'][sensor_array[sensor[i]]['labeled_anomaly']], 'mo', mfc='none', label='technician labeled anomalies')
plt.plot(detections_array[i]['prediction'][detections_array[i]['anomaly']], 'r+', label='machine detected anomalies')
plt.legend()
plt.ylabel(sensor[i])
plt.show()
print("\n LSTM script end.")
| 38.907407 | 194 | 0.671857 | 2,034 | 14,707 | 4.66765 | 0.117502 | 0.043817 | 0.025068 | 0.024647 | 0.821045 | 0.811671 | 0.793448 | 0.78102 | 0.78102 | 0.78102 | 0 | 0.015823 | 0.114775 | 14,707 | 377 | 195 | 39.01061 | 0.713419 | 0.160672 | 0 | 0.740458 | 0 | 0 | 0.156203 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.038168 | 0 | 0.038168 | 0.389313 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
357c732ea7136d2889cda027fe0436f8645a4363 | 1,977 | py | Python | src/viola/testing.py | mmiki21/Viola-SV | 10fe2c326749024551d3e41f900b37e20dbde35c | [
"Apache-2.0"
] | 13 | 2021-05-03T18:43:25.000Z | 2022-03-26T04:57:01.000Z | src/viola/testing.py | mmiki21/Viola-SV | 10fe2c326749024551d3e41f900b37e20dbde35c | [
"Apache-2.0"
] | 6 | 2021-04-02T08:04:23.000Z | 2022-03-24T11:34:43.000Z | src/viola/testing.py | mmiki21/Viola-SV | 10fe2c326749024551d3e41f900b37e20dbde35c | [
"Apache-2.0"
] | 3 | 2022-02-23T03:36:12.000Z | 2022-03-07T07:05:00.000Z | import viola
from viola.core.bedpe import Bedpe
from viola.core.vcf import Vcf
import pandas as pd
def assert_bedpe_equal(left, right, check_like=True):
ls_left_table = left.table_list
ls_right_table = right.table_list
assert len(ls_left_table) == len(ls_right_table)
assert set(ls_left_table) == set(ls_right_table)
for tablename in ls_left_table:
df_left = left.get_table(tablename)
df_right = right.get_table(tablename)
id_left = df_left.columns[0]
id_right = df_right.columns[0]
df_left.sort_values(id_left, inplace=True, ignore_index=True)
df_right.sort_values(id_right, inplace=True, ignore_index=True)
if df_left.empty & df_right.empty:
continue
try:
pd.testing.assert_frame_equal(df_left, df_right, check_like=check_like, check_exact=True)
except AssertionError:
print('\nwhen asserting {} table, following error occured!'.format(tablename))
raise
def assert_vcf_equal(left, right, check_like=True):
ls_left_table = left.table_list
ls_right_table = right.table_list
assert len(ls_left_table) == len(ls_right_table)
assert set(ls_left_table) == set(ls_right_table)
for tablename in ls_left_table:
df_left = left.get_table(tablename)
df_right = right.get_table(tablename)
id_left = df_left.columns[0]
id_right = df_right.columns[0]
df_left.sort_values(id_left, inplace=True, ignore_index=True)
df_right.sort_values(id_right, inplace=True, ignore_index=True)
if df_left.empty & df_right.empty:
continue
try:
pd.testing.assert_frame_equal(df_left, df_right, check_like=check_like, check_exact=True)
except AssertionError:
print('\nwhen asserting {} table, following error occured!'.format(tablename))
raise
# assert metadata
assert left._metadata == right._metadata
| 39.54 | 101 | 0.687911 | 281 | 1,977 | 4.519573 | 0.188612 | 0.070866 | 0.069291 | 0.069291 | 0.880315 | 0.880315 | 0.880315 | 0.880315 | 0.880315 | 0.880315 | 0 | 0.002616 | 0.226606 | 1,977 | 49 | 102 | 40.346939 | 0.827992 | 0.007587 | 0 | 0.837209 | 0 | 0 | 0.052041 | 0 | 0 | 0 | 0 | 0 | 0.302326 | 1 | 0.046512 | false | 0 | 0.093023 | 0 | 0.139535 | 0.046512 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
3595b042857ebc343975566768abd9ecefaa9c77 | 3,223 | py | Python | tests/test_deque.py | EndreoT/DataStructures | 3359c142a3a3b1eb2e275a90544973dd709b0d29 | [
"MIT"
] | null | null | null | tests/test_deque.py | EndreoT/DataStructures | 3359c142a3a3b1eb2e275a90544973dd709b0d29 | [
"MIT"
] | null | null | null | tests/test_deque.py | EndreoT/DataStructures | 3359c142a3a3b1eb2e275a90544973dd709b0d29 | [
"MIT"
] | null | null | null | import unittest
from deque import Deque
from node import Node
class TestQueue(unittest.TestCase):
def setUp(self):
self.q = Deque()
def test_empty_deque(self):
self.assertEqual(self.q.get_front(), None)
self.assertEqual(self.q.get_rear(), None)
self.assertEqual(self.q.get_front(), self.q.get_rear())
self.assertEqual(self.q.get_size(), 0)
self.assertEqual(self.q.remove_front(), None)
def test_one_element_deque(self):
q = Deque(1)
self.assertEqual(q.get_front().get_value(), 1)
self.assertEqual(q.get_rear().get_value(), 1)
self.assertEqual(type(q.get_front()), Node)
self.assertEqual(type(q.get_rear()), Node)
self.assertEqual(q.get_rear(), q.get_front())
self.assertEqual(q.get_rear().get_previous(), None)
self.assertEqual(q.get_rear().get_next(), None)
self.assertEqual(q.get_front().get_next(), None)
self.assertEqual(q.get_front().get_previous(), None)
self.assertEqual(q.get_size(), 1)
def test_add_front(self):
for i in range(1, 4):
self.q.add_front(i)
self.assertEqual(self.q.get_front().get_value(), 3)
self.assertEqual(self.q.get_rear().get_next().get_value(), 2)
self.assertEqual(self.q.get_front().get_next(), None)
self.assertEqual(self.q.get_front().get_previous().get_value(), 2)
self.assertEqual(self.q.get_size(), 3)
def test_add_rear(self):
for i in range(1, 4):
self.q.add_rear(i)
self.assertEqual(self.q.get_front().get_value(), 1)
self.assertEqual(self.q.get_rear().get_next().get_value(), 2)
self.assertEqual(self.q.get_front().get_next(), None)
self.assertEqual(self.q.get_front().get_previous().get_value(), 2)
self.assertEqual(self.q.get_size(), 3)
def test_remove_front(self):
for i in range(1, 8):
self.q.add_front(i)
removed = self.q.remove_front()
self.assertEqual(self.q.get_size(), 6)
self.assertEqual(self.q.get_front().get_value(), 6)
self.assertEqual(self.q.get_front().get_next(), None)
self.assertEqual(self.q.get_front().get_previous().get_value(), 5)
self.assertEqual(removed.get_value(), 7)
self.assertEqual(removed.get_previous(), None)
def test_remove_rear(self):
for i in range(1, 8):
self.q.add_rear(i)
removed = self.q.remove_rear()
self.assertEqual(self.q.get_size(), 6)
self.assertEqual(self.q.get_rear().get_value(), 6)
self.assertEqual(self.q.get_rear().get_previous(), None)
self.assertEqual(self.q.get_rear().get_next().get_value(), 5)
self.assertEqual(removed.get_value(), 7)
self.assertEqual(removed.get_previous(), None)
def test_add_both(self):
self.q.add_front(2)
self.q.add_rear(1)
self.assertEqual(self.q.get_rear().get_value(), 1)
self.assertEqual(self.q.get_front().get_value(), 2)
self.assertEqual(self.q.get_rear().get_next().get_value(), 2)
self.assertEqual(self.q.get_front().get_previous().get_value(), 1)
if __name__ == "__main__":
unittest.main()
| 37.917647 | 74 | 0.635743 | 472 | 3,223 | 4.116525 | 0.097458 | 0.316521 | 0.264025 | 0.277921 | 0.833762 | 0.776634 | 0.744725 | 0.66701 | 0.539887 | 0.464231 | 0 | 0.013656 | 0.204778 | 3,223 | 84 | 75 | 38.369048 | 0.74444 | 0 | 0 | 0.347826 | 0 | 0 | 0.002482 | 0 | 0 | 0 | 0 | 0 | 0.594203 | 1 | 0.115942 | false | 0 | 0.043478 | 0 | 0.173913 | 0 | 0 | 0 | 0 | null | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
57a56459a1620898f667910fbc474dac679355b2 | 23,983 | py | Python | pycqed/measurement/waveform_control_CC/QWG_fluxing_seqs.py | nuttamas/PycQED_py3 | 1ee35c7428d36ed42ba4afb5d4bda98140b2283e | [
"MIT"
] | 60 | 2016-08-03T10:00:18.000Z | 2021-11-10T11:46:16.000Z | pycqed/measurement/waveform_control_CC/QWG_fluxing_seqs.py | nuttamas/PycQED_py3 | 1ee35c7428d36ed42ba4afb5d4bda98140b2283e | [
"MIT"
] | 512 | 2016-08-03T17:10:02.000Z | 2022-03-31T14:03:43.000Z | pycqed/measurement/waveform_control_CC/QWG_fluxing_seqs.py | nuttamas/PycQED_py3 | 1ee35c7428d36ed42ba4afb5d4bda98140b2283e | [
"MIT"
] | 34 | 2016-10-19T12:00:52.000Z | 2022-03-19T04:43:26.000Z | from pycqed.utilities.general import mopen
from os.path import join, dirname
import numpy as np
base_qasm_path = join(dirname(__file__), 'qasm_files')
from pycqed.measurement.waveform_control_CC.multi_qubit_qasm_seqs \
import cal_points_2Q
def ramZ_flux_latency(q0_name, wait_after_flux=20):
"""
Sequence designed to calibrate the delay between the
QWG_trigger and the start of the flux pulse
Consists of a single point. Intended to change the latency parameter
in the configuration that is used in compilation.
"""
filename = join(base_qasm_path, 'RamZ_latency_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \n'.format(q0_name))
# simultaneous MW and flux pulse
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('X90 {} \n'.format(q0_name))
qasm_file.writelines('square {}\n'.format(q0_name))
qasm_file.writelines('I {}\n'.format(wait_after_flux))
qasm_file.writelines('X90 {}\n'.format(q0_name))
qasm_file.writelines('RO {} \n'.format(q0_name))
qasm_file.close()
return qasm_file
def chevron_block_seq(q0_name, q1_name, no_of_points,
excite_q1=False, wait_after_trigger=40e-9,
wait_during_flux=400e-9, clock_cycle=1e-9,
RO_target='all', mw_pulse_duration=40e-9,
cal_points=True):
'''
N.B. this sequence has been edited for compatibility with the XFU compiler
Sequence for measuring a block of a chevron, i.e. using different codewords
for different pulse lengths.
Args:
q0, q1 (str): names of the addressed qubits.
q0 is the pulse that experiences the flux pulse.
RO_target (str): can be q0, q1, or 'all'
excite_q1 (bool): choose whether to excite q1, thus choosing
between the |01> <-> |10> and the |11> <-> |20>
swap.
wait_after_trigger (float): delay time in seconds after sending the
trigger for the flux pulse
clock_cycle (float): period of the internal AWG clock
wait_time (int): wait time between triggering QWG and RO
cal_points (bool): whether to use calibration points or not
'''
filename = join(base_qasm_path, 'chevron_block_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0_name, q1_name))
for i in range(no_of_points):
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('QWG_trigger_{} {}\n'.format(i, q0_name))
if excite_q1:
wait_after_trigger -= mw_pulse_duration
qasm_file.writelines('X180 {}\n'.format(q0_name))
if excite_q1:
qasm_file.writelines('X180 {}\n'.format(q1_name))
qasm_file.writelines('CZ {} {}\n'.format(q0_name, q1_name))
if excite_q1:
# q0 is rotated to ground-state to have better contrast
# (|0> and |2> instead of |1> and |2>)
qasm_file.writelines('X180 {}\n'.format(q0_name))
if RO_target == 'all':
qasm_file.writelines('RO {} | RO {} \n'.format(q0_name, q1_name))
else:
qasm_file.writelines('RO {} \n'.format(RO_target))
if cal_points:
# Add calibration pulses
cal_pulses = []
for seq in cal_points_2Q:
cal_pulses += [[seq[0], seq[1], 'RO ' + RO_target + '\n']]
qasm_file.close()
return qasm_file
def SWAPN(q0_name, q1_name, nr_pulses: list,
excite_q1=False,
RO_target='all',
cal_points=True):
'''
Args:
q0, q1 (str): names of the addressed qubits.
q0 is the pulse that experiences the flux pulse.
RO_target (str): can be q0, q1, or 'all'
excite_q1 (bool): choose whether to excite q1, thus choosing
between the |01> <-> |10> and the |11> <-> |20>
swap.
cal_points (bool): whether to use calibration points or not
'''
filename = join(base_qasm_path, 'chevron_block_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0_name, q1_name))
for i, N in enumerate(nr_pulses):
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('QWG_trigger_{} {}\n'.format(i, q0_name))
qasm_file.writelines('X180 {}\n'.format(q0_name))
if excite_q1:
qasm_file.writelines('X180 {}\n'.format(q1_name))
for n in range(N):
qasm_file.writelines('square {} \n'.format(q0_name))
if excite_q1:
# q0 is rotated to ground-state to have better contrast
# (|0> and |2> instead of |1> and |2>)
qasm_file.writelines('X180 {}\n'.format(q0_name))
qasm_file.writelines('RO {} \n'.format(RO_target))
if cal_points:
# Add calibration pulses
cal_pulses = []
for seq in cal_points_2Q:
cal_pulses += [[seq[0].format(q0_name) +
seq[1].format(q1_name) +
'RO {} \n'.format(RO_target)]]
for seq in cal_pulses:
qasm_file.writelines('\ninit_all\n')
for p in seq:
qasm_file.writelines(p)
qasm_file.close()
return qasm_file
def CZ_calibration_seq(q0, q1, RO_target='all',
vary_single_q_phase=True,
cases=('no_excitation', 'excitation')):
'''
Sequence used to calibrate flux pulses for CZ gates.
Timing of the sequence:
q0: X90 C-Phase Rphi90 RO
q1: (X180) -- (X180) RO
Args:
q0, q1 (str): names of the addressed qubits
RO_target (str): can be q0, q1, or 'all'
excitations (bool/str): can be True, False, or 'both_cases'
'''
filename = join(base_qasm_path, 'CZ_calibration_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0, q1))
for case in cases:
qasm_file.writelines('\ninit_all\n')
if case == 'excitation':
qasm_file.writelines('X180 {} | '.format(q1))
qasm_file.writelines('X90 {}\n'.format(q0))
# temporary workaround to deal with limitation in the QASM config
# qasm_file.writelines('CZ {} \n'.format(q0))
qasm_file.writelines('CZ {} {}\n'.format(q0, q1))
if case == 'excitation':
qasm_file.writelines('X180 {} | '.format(q1))
if vary_single_q_phase:
qasm_file.writelines('Rphi90 {}\n'.format(q0))
else:
qasm_file.writelines('mX90 {}\n'.format(q0))
if 'RO_target' == 'all':
qasm_file.writelines('RO {} | RO {} \n'.format(q0, q1))
else:
qasm_file.writelines('RO {} \n'.format(RO_target))
qasm_file.close()
return qasm_file
def two_qubit_tomo_bell(bell_state, q0, q1, RO_target='all'):
'''
Two qubit bell state tomography.
Args:
bell_state (int): index of prepared bell state
0 : |00>-|11>
1 : |00>+|11>
2 : |01>-|10>
3 : |01>+|10>
q0, q1 (str): names of the target qubits
RO_target (str): can be q0, q1, or 'all'
'''
if RO_target == 'all':
# This is a bit of a hack as RO all qubits is the same instruction
# as any specific qubit
RO_target = q0
tomo_pulses = ['I ', 'X180 ', 'Y90 ', 'mY90 ', 'X90 ', 'mX90 ']
tomo_list_q0 = []
tomo_list_q1 = []
for tp in tomo_pulses:
tomo_list_q0 += [tp + q0]
tomo_list_q1 += [tp + q1]
tomo_list_q0[0] = 'I {}'.format(q0)
tomo_list_q1[0] = 'I {}'.format(q1)
# Choose a bell state and set the corresponding preparation pulses
if bell_state % 10 == 0: # |Phi_m>=|00>-|11>
prep_pulse_q0 = 'Y90 {}'.format(q0)
prep_pulse_q1 = 'Y90 {}'.format(q1)
elif bell_state % 10 == 1: # |Phi_p>=|00>+|11>
prep_pulse_q0 = 'mY90 {}'.format(q0)
prep_pulse_q1 = 'Y90 {}'.format(q1)
elif bell_state % 10 == 2: # |Psi_m>=|01>-|10>
prep_pulse_q0 = 'Y90 {}'.format(q0)
prep_pulse_q1 = 'mY90 {}'.format(q1)
elif bell_state % 10 == 3: # |Psi_p>=|01>+|10>
prep_pulse_q0 = 'mY90 {}'.format(q0)
prep_pulse_q1 = 'mY90 {}'.format(q1)
else:
raise ValueError('Bell state {} is not defined.'.format(bell_state))
after_pulse = 'mY90 {}\n'.format(q1)
# Disable preparation pulse on one or the other qubit for debugging
if bell_state//10 == 1:
prep_pulse_q1 = 'I {}'.format(q0)
elif bell_state//10 == 2:
prep_pulse_q0 = 'I {}'.format(q1)
# Write tomo sequence
filename = join(base_qasm_path, 'two_qubit_tomo_bell.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0, q1))
for p_q1 in tomo_list_q1:
for p_q0 in tomo_list_q0:
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('{} | {} \n'.format(prep_pulse_q0,
prep_pulse_q1))
qasm_file.writelines('CZ {} {} \n'.format(q0, q1))
qasm_file.writelines(after_pulse)
qasm_file.writelines('{} | {}\n'.format(p_q1, p_q0))
qasm_file.writelines('RO ' + RO_target + ' \n')
# Add calibration pulses
cal_pulses = []
# every calibration point is repeated 7 times. This is copied from the
# script for Tektronix driven qubits. I do not know if this repetition
# is important or even necessary here.
for seq in cal_points_2Q:
cal_pulses += [[seq[0].format(q0), seq[1].format(q1),
'RO ' + RO_target + '\n']] * 7
for seq in cal_pulses:
qasm_file.writelines('\ninit_all\n')
for p in seq:
qasm_file.writelines(p)
qasm_file.close()
return qasm_file
def grover_seq(q0_name, q1_name, RO_target='all',
precompiled_flux=True, cal_points: bool=True):
'''
Writes the QASM sequence for Grover's algorithm on two qubits.
Sequence:
q0: G0 - - mY90 - - mY90 - RO
CZ_ij CZ
q1: G1 - - mY90 - - mY90 - RO
whit all combinations of (ij) = omega.
G0 and G1 are Y90 or Y90, depending on the (ij).
Args:
q0_name, q1_name (string):
Names of the qubits to which the sequence is applied.
RO_target (string):
Readout target. Can be a qubit name or 'all'.
precompiled_flux (bool):
Determies if the full waveform for the flux pulses is
precompiled, thus only needing one trigger at the start,
or if every flux pulse should be triggered individually.
cal_points (bool):
Whether to add calibration points.
Returns:
qasm_file: a reference to the new QASM file object.
'''
if not precompiled_flux:
raise NotImplementedError('Currently only precompiled flux pulses '
'are supported.')
filename = join(base_qasm_path, 'Grover_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \n'.format(q0_name))
qasm_file.writelines('qubit {} \n'.format(q1_name))
if RO_target == 'all':
RO_line = 'RO {} | RO {}\n'.format(q0_name, q1_name)
else:
RO_line = 'RO {} \n'.format(RO_target)
for G0 in ['Y90', 'mY90']:
for G1 in ['Y90', 'mY90']:
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('{} {} | {} {}\n'.format(G0, q0_name,
G1, q1_name))
qasm_file.writelines('grover_CZ {} {}\n'.format(q0_name, q1_name))
qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name,
q1_name))
qasm_file.writelines('cz {} {}\n'.format(q0_name, q1_name))
qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name,
q1_name))
qasm_file.writelines(RO_line)
# Add calibration points
if cal_points:
cal_pulses = []
for seq in cal_points_2Q:
cal_pulses += [[seq[0].format(q0_name), seq[1].format(q1_name),
RO_line]]
for seq in cal_pulses:
qasm_file.writelines('\ninit_all\n')
for p in seq:
qasm_file.writelines(p)
qasm_file.close()
return qasm_file
def grover_tomo_seq(q0_name, q1_name, omega, RO_target='all',
precompiled_flux=True):
'''
Writes the QASM sequence to take a state tomography of the output state
of Grover's algorithm on two qubits.
Sequence:
q0: G0 - - mY90 - - mY90 - RO
CZ_ij CZ
q1: G1 - - mY90 - - mY90 - RO
where (ij) is the binary representation of omega.
G0 and G1 are Y90 or Y90, depending on the (ij).
Args:
q0_name, q1_name (string):
Names of the qubits to which the sequence is applied.
omega (int):
Deterines which (ij) for the CZ_ij.
RO_target (string):
Readout target. Can be a qubit name or 'all'.
precompiled_flux (bool):
Determies if the full waveform for the flux pulses is
precompiled, thus only needing one trigger at the start,
or if every flux pulse should be triggered individually.
Returns:
qasm_file: a reference to the new QASM file object.
'''
if not precompiled_flux:
raise NotImplementedError('Currently only precompiled flux pulses '
'are supported.')
tomo_pulses = ['I ', 'X180 ', 'Y90 ', 'mY90 ', 'X90 ', 'mX90 ']
tomo_list_q0 = []
tomo_list_q1 = []
for tp in tomo_pulses:
tomo_list_q0 += [tp + q0_name]
tomo_list_q1 += [tp + q1_name]
if omega == 0:
G0 = 'Y90'
G1 = 'Y90'
elif omega == 1:
G0 = 'Y90'
G1 = 'mY90'
elif omega == 2:
G0 = 'mY90'
G1 = 'Y90'
elif omega == 3:
G0 = 'mY90'
G1 = 'mY90'
else:
raise ValueError('omega must be in [0, 3]')
if RO_target == 'all':
RO_line = 'RO {} | RO {}\n'.format(q0_name, q1_name)
else:
RO_line = 'RO {} \n'.format(RO_target)
filename = join(base_qasm_path, 'Grover_tomo_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \n'.format(q0_name))
qasm_file.writelines('qubit {} \n'.format(q1_name))
for p_q1 in tomo_list_q1:
for p_q0 in tomo_list_q0:
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('{} {} | {} {}\n'.format(G0, q0_name,
G1, q1_name))
qasm_file.writelines('grover_CZ {} {}\n'.format(q0_name, q1_name))
qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name,
q1_name))
qasm_file.writelines('cz {} {}\n'.format(q0_name, q1_name))
qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name,
q1_name))
qasm_file.writelines('{} | {}\n'.format(p_q1, p_q0))
qasm_file.writelines(RO_line)
# Add calibration pulses
cal_pulses = []
# every calibration point is repeated 7 times. This is copied from the
# script for Tektronix driven qubits. I do not know if this repetition
# is important or even necessary here.
for seq in cal_points_2Q:
cal_pulses += [[seq[0].format(q0_name), seq[1].format(q1_name),
RO_line]] * 7
for seq in cal_pulses:
qasm_file.writelines('\ninit_all\n')
for p in seq:
qasm_file.writelines(p)
qasm_file.close()
return qasm_file
def grover_test_seq(q0_name, q1_name, RO_target='all',
precompiled_flux=True, cal_points: bool=True):
'''
Writes the QASM sequence for Grover's algorithm on two qubits.
Sequence:
q0: G0 - - mY90 - - mY90 - RO
CZ_ij CZ
q1: G1 - - mY90 - - mY90 - RO
whit all combinations of (ij) = omega.
G0 and G1 are Y90 or Y90, depending on the (ij).
Args:
q0_name, q1_name (string):
Names of the qubits to which the sequence is applied.
RO_target (string):
Readout target. Can be a qubit name or 'all'.
precompiled_flux (bool):
Determies if the full waveform for the flux pulses is
precompiled, thus only needing one trigger at the start,
or if every flux pulse should be triggered individually.
cal_points (bool):
Whether to add calibration points.
Returns:
qasm_file: a reference to the new QASM file object.
'''
if not precompiled_flux:
raise NotImplementedError('Currently only precompiled flux pulses '
'are supported.')
filename = join(base_qasm_path, 'Grover_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \n'.format(q0_name))
qasm_file.writelines('qubit {} \n'.format(q1_name))
if RO_target == 'all':
RO_line = 'RO {} | RO {}\n'.format(q0_name, q1_name)
else:
RO_line = 'RO {} \n'.format(RO_target)
for G0 in ['Y90', 'mY90']:
for G1 in ['Y90', 'mY90']:
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('{} {} | {} {}\n'.format(G0, q0_name,
G1, q1_name))
qasm_file.writelines('grover_CZ {} {}\n'.format(q0_name, q1_name))
qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name,
q1_name))
# qasm_file.writelines('cz {} {}\n'.format(q0_name, q1_name))
# qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name,
# q1_name))
qasm_file.writelines(RO_line)
# Add calibration points
if cal_points:
cal_pulses = []
for seq in cal_points_2Q:
cal_pulses += [[seq[0].format(q0_name), seq[1].format(q1_name),
RO_line]]
for seq in cal_pulses:
qasm_file.writelines('\ninit_all\n')
for p in seq:
qasm_file.writelines(p)
qasm_file.close()
return qasm_file
def grover_test_tomo_seq(q0_name, q1_name, omega, RO_target='all',
precompiled_flux=True):
'''
Test sequence to debug Grover's algorithm.
'''
if not precompiled_flux:
raise NotImplementedError('Currently only precompiled flux pulses '
'are supported.')
tomo_pulses = ['I ', 'X180 ', 'Y90 ', 'mY90 ', 'X90 ', 'mX90 ']
tomo_list_q0 = []
tomo_list_q1 = []
for tp in tomo_pulses:
tomo_list_q0 += [tp + q0_name]
tomo_list_q1 += [tp + q1_name]
if omega == 0:
G0 = 'Y90'
G1 = 'Y90'
elif omega == 1:
G0 = 'Y90'
G1 = 'mY90'
elif omega == 2:
G0 = 'mY90'
G1 = 'Y90'
elif omega == 3:
G0 = 'mY90'
G1 = 'mY90'
else:
raise ValueError('omega must be in [0, 3]')
if RO_target == 'all':
RO_line = 'RO {} | RO {}\n'.format(q0_name, q1_name)
else:
RO_line = 'RO {} \n'.format(RO_target)
filename = join(base_qasm_path, 'Grover_tomo_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \n'.format(q0_name))
qasm_file.writelines('qubit {} \n'.format(q1_name))
for p_q1 in tomo_list_q1:
for p_q0 in tomo_list_q0:
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('{} {} | {} {}\n'.format(G0, q0_name,
G1, q1_name))
qasm_file.writelines('grover_cz {} {}\n'.format(q0_name, q1_name))
qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name,
q1_name))
qasm_file.writelines('cz {} {}\n'.format(q0_name, q1_name))
# qasm_file.writelines('Y90 {} | Y90 {}\n'.format(q0_name,
# q1_name))
qasm_file.writelines('{} | {}\n'.format(p_q1, p_q0))
qasm_file.writelines(RO_line)
# Add calibration pulses
cal_pulses = []
# every calibration point is repeated 7 times. This is copied from the
# script for Tektronix driven qubits. I do not know if this repetition
# is important or even necessary here.
for seq in cal_points_2Q:
cal_pulses += [[seq[0].format(q0_name), seq[1].format(q1_name),
RO_line]] * 7
for seq in cal_pulses:
qasm_file.writelines('\ninit_all\n')
for p in seq:
qasm_file.writelines(p)
qasm_file.close()
return qasm_file
def purity_CZ_seq(q0, q1, RO_target='all'):
"""
Creates the |00> + |11> Bell state and does a partial tomography in
order to determine the purity of both qubits.
"""
filename = join(base_qasm_path, 'purity_CZ_seq.qasm')
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0, q1))
tomo_list = ['mX90', 'mY90', 'I']
for p_pulse in tomo_list:
# Create a Bell state: |00> + |11>
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('mY90 {} | Y90 {} \n'.format(q0, q1))
qasm_file.writelines('CZ {} {} \n'.format(q0, q1))
qasm_file.writelines('mY90 {}\n'.format(q1))
# Perform pulses to measure the purity of both qubits
qasm_file.writelines('{} {} | {} {}\n'.format(p_pulse, q0,
p_pulse, q1))
if RO_target == 'all':
qasm_file.writelines('RO {} | RO {} \n'.format(q0, q1))
else:
qasm_file.writelines('RO {} \n'.format(RO_target))
qasm_file.close()
return qasm_file
def purity_N_CZ_seq(q0: str, q1: str, N: int, RO_target: str='all'):
"""
Creates the |00> + |11> Bell state and does a partial tomography in
order to determine the purity of both qubits.
"""
filename = join(base_qasm_path, 'purity_{}_CZ_seq.qasm'.format(N))
qasm_file = mopen(filename, mode='w')
qasm_file.writelines('qubit {} \nqubit {} \n'.format(q0, q1))
tomo_list = ['mX90', 'mY90', 'I']
for i, p_pulse in enumerate(tomo_list):
# Create a Bell state: |00> + |11>
qasm_file.writelines('\ninit_all\n')
qasm_file.writelines('qwg_trigger_{} {}\n'.format(0, q0))
qasm_file.writelines('mY90 {} | Y90 {} \n'.format(q0, q1))
for n in range(N):
qasm_file.writelines('dummy_CZ {} {} \n'.format(q0, q1))
qasm_file.writelines('mY90 {}\n'.format(q1))
# Perform pulses to measure the purity of both qubits
qasm_file.writelines('{} {} | {} {}\n'.format(p_pulse, q0,
p_pulse, q1))
if RO_target == 'all':
qasm_file.writelines('RO {} | RO {} \n'.format(q0, q1))
else:
qasm_file.writelines('RO {} \n'.format(RO_target))
qasm_file.close()
return qasm_file
| 37.356698 | 79 | 0.556686 | 3,139 | 23,983 | 4.052246 | 0.090156 | 0.093082 | 0.154245 | 0.038836 | 0.824135 | 0.809827 | 0.800943 | 0.789701 | 0.777044 | 0.752594 | 0 | 0.041887 | 0.322103 | 23,983 | 641 | 80 | 37.414977 | 0.740497 | 0.279365 | 0 | 0.805405 | 0 | 0 | 0.132134 | 0.007974 | 0 | 0 | 0 | 0 | 0 | 1 | 0.02973 | false | 0 | 0.010811 | 0 | 0.07027 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
57e59a41b444f8b1198e53a31beb8895923cd651 | 1,175 | py | Python | app/map/forms/__init__.py | jg-725/IS219-FlaskAppProject | 316aa298eda1bcda766ed085bb6f26ca7da7dfee | [
"BSD-3-Clause"
] | null | null | null | app/map/forms/__init__.py | jg-725/IS219-FlaskAppProject | 316aa298eda1bcda766ed085bb6f26ca7da7dfee | [
"BSD-3-Clause"
] | null | null | null | app/map/forms/__init__.py | jg-725/IS219-FlaskAppProject | 316aa298eda1bcda766ed085bb6f26ca7da7dfee | [
"BSD-3-Clause"
] | null | null | null | from flask_wtf import FlaskForm
from wtforms import validators
from wtforms.fields import *
class csv_upload(FlaskForm):
file = FileField()
submit = SubmitField()
class create_location_form(FlaskForm):
title = StringField('City Name', description="Add The City Name")
latitude = StringField('Latitude:', [
validators.DataRequired(),
], description="Positive Value")
longitude = StringField('Longitude:', [
validators.DataRequired(),
], description="Negative Value")
population = IntegerField('Population:', [
validators.DataRequired(),
], description="Add Population Value ")
submit = SubmitField()
class edit_location_form(FlaskForm):
title = StringField('City Name', description="Edit The City Name")
latitude = StringField('Latitude:', [
validators.DataRequired(),
], description=" Positive Value ")
longitude = StringField('Longitude:', [
validators.DataRequired(),
], description="Negative Value ")
population = IntegerField('Population:', [
validators.DataRequired(),
], description="Add Population Value ")
submit = SubmitField() | 23.979592 | 70 | 0.674894 | 103 | 1,175 | 7.640777 | 0.330097 | 0.167726 | 0.251588 | 0.066074 | 0.800508 | 0.800508 | 0.800508 | 0.800508 | 0.658196 | 0.658196 | 0 | 0 | 0.203404 | 1,175 | 49 | 71 | 23.979592 | 0.840812 | 0 | 0 | 0.7 | 0 | 0 | 0.181973 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.1 | 0 | 0.6 | 0 | 0 | 0 | 0 | null | 0 | 1 | 0 | 1 | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 7 |
17b57f56133daabf92fa3b1bc58aa9bfbe85e35a | 4,137 | py | Python | chsdi/models/vector/kogis.py | procrastinatio/mf-chsdi3 | 206e9362860376338cc83ded902790dc216e2563 | [
"BSD-3-Clause"
] | null | null | null | chsdi/models/vector/kogis.py | procrastinatio/mf-chsdi3 | 206e9362860376338cc83ded902790dc216e2563 | [
"BSD-3-Clause"
] | null | null | null | chsdi/models/vector/kogis.py | procrastinatio/mf-chsdi3 | 206e9362860376338cc83ded902790dc216e2563 | [
"BSD-3-Clause"
] | null | null | null | # -*- coding: utf-8 -*-
from sqlalchemy import Column, Integer
from sqlalchemy.types import Numeric, Unicode
from chsdi.models import register, bases
from chsdi.models.vector import Vector, Geometry2D
Base = bases['kogis']
class Agnes(Base, Vector):
__tablename__ = 'agnes'
__table_args__ = ({'schema': 'fpds', 'autoload': False})
__template__ = 'templates/htmlpopup/agnes.mako'
__bodId__ = 'ch.swisstopo.fixpunkte-agnes'
__label__ = 'id'
id = Column('no', Unicode, primary_key=True)
url = Column('url', Unicode)
bgdi_id = Column('bgdi_id', Integer)
the_geom = Column(Geometry2D)
register('ch.swisstopo.fixpunkte-agnes', Agnes)
class FixpunkteLfp1(Base, Vector):
__tablename__ = 'punkt_lage_lfp1'
__table_args__ = ({'schema': 'fpds', 'autoload': False})
__template__ = 'templates/htmlpopup/fixpunkte.mako'
__bodId__ = 'ch.swisstopo.fixpunkte-lfp1'
__label__ = 'id'
id = Column('pointid', Unicode, primary_key=True)
punktname = Column('punktname', Unicode)
nummer = Column('nummer', Unicode)
status = Column('status', Unicode)
nbident = Column('nbident', Unicode)
x03 = Column('x03', Numeric)
y03 = Column('y03', Numeric)
n95 = Column('n95', Numeric)
e95 = Column('e95', Numeric)
h02 = Column('h02', Numeric)
zugang = Column('zugang', Unicode)
url = Column('url', Unicode)
bgdi_created = Column('bgdi_created', Unicode)
the_geom = Column(Geometry2D)
register('ch.swisstopo.fixpunkte-lfp1', FixpunkteLfp1)
class FixpunkteLfp2(Base, Vector):
__tablename__ = 'punkt_lage_lfp2'
__table_args__ = ({'schema': 'fpds', 'autoload': False})
__template__ = 'templates/htmlpopup/fixpunkte.mako'
__bodId__ = 'ch.swisstopo.fixpunkte-lfp2'
__label__ = 'id'
id = Column('pointid', Unicode, primary_key=True)
nbident = Column('nbident', Unicode)
punktname = Column('punktname', Unicode)
status = Column('status', Unicode)
nummer = Column('nummer', Unicode)
x03 = Column('x03', Numeric)
y03 = Column('y03', Numeric)
n95 = Column('n95', Numeric)
e95 = Column('e95', Numeric)
h02 = Column('h02', Numeric)
zugang = Column('zugang', Unicode)
url = Column('url', Unicode)
bgdi_created = Column('bgdi_created', Unicode)
the_geom = Column(Geometry2D)
register('ch.swisstopo.fixpunkte-lfp2', FixpunkteLfp2)
class FixpunkteHfp1(Base, Vector):
__tablename__ = 'punkt_hoehe_hfp1'
__table_args__ = ({'schema': 'fpds', 'autoload': True})
__template__ = 'templates/htmlpopup/fixpunkte.mako'
__bodId__ = 'ch.swisstopo.fixpunkte-hfp1'
__label__ = 'id'
id = Column('pointid', Unicode, primary_key=True)
bgdi_label = Column('bgdi_label', Unicode)
nbident = Column('nbident', Unicode)
punktname = Column('punktname', Unicode)
status = Column('status', Unicode)
nummer = Column('nummer', Unicode)
x03 = Column('x03', Numeric)
y03 = Column('y03', Numeric)
n95 = Column('n95', Numeric)
e95 = Column('e95', Numeric)
h02 = Column('h02', Numeric)
zugang = Column('zugang', Unicode)
url = Column('url', Unicode)
bgdi_created = Column('bgdi_created', Unicode)
the_geom = Column(Geometry2D)
register('ch.swisstopo.fixpunkte-hfp1', FixpunkteHfp1)
class FixpunkteHfp2(Base, Vector):
__tablename__ = 'punkt_hoehe_hfp2'
__table_args__ = ({'schema': 'fpds', 'autoload': True})
__template__ = 'templates/htmlpopup/fixpunkte.mako'
__bodId__ = 'ch.swisstopo.fixpunkte-hfp2'
__label__ = 'id'
id = Column('pointid', Unicode, primary_key=True)
nbident = Column('nbident', Unicode)
punktname = Column('punktname', Unicode)
status = Column('status', Unicode)
nummer = Column('nummer', Unicode)
x03 = Column('x03', Numeric)
y03 = Column('y03', Numeric)
n95 = Column('n95', Numeric)
e95 = Column('e95', Numeric)
h02 = Column('h02', Numeric)
zugang = Column('zugang', Unicode)
url = Column('url', Unicode)
bgdi_created = Column('bgdi_created', Unicode)
the_geom = Column(Geometry2D)
register('ch.swisstopo.fixpunkte-hfp2', FixpunkteHfp2)
| 34.190083 | 60 | 0.673193 | 459 | 4,137 | 5.77342 | 0.150327 | 0.041509 | 0.075472 | 0.035849 | 0.809057 | 0.714717 | 0.714717 | 0.714717 | 0.695472 | 0.641132 | 0 | 0.031582 | 0.181049 | 4,137 | 120 | 61 | 34.475 | 0.75059 | 0.005076 | 0 | 0.712871 | 0 | 0 | 0.222654 | 0.106466 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.039604 | 0 | 0.940594 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 7 |
aa36a6fe7866049bd9b571b53e77325c1cd01652 | 741 | py | Python | stubs/micropython-v1_13-95-pyboard/utime.py | mattytrentini/micropython-stubs | 4d596273823b69e9e5bcf5fa67f249c374ee0bbc | [
"MIT"
] | null | null | null | stubs/micropython-v1_13-95-pyboard/utime.py | mattytrentini/micropython-stubs | 4d596273823b69e9e5bcf5fa67f249c374ee0bbc | [
"MIT"
] | null | null | null | stubs/micropython-v1_13-95-pyboard/utime.py | mattytrentini/micropython-stubs | 4d596273823b69e9e5bcf5fa67f249c374ee0bbc | [
"MIT"
] | null | null | null | """
Module: 'utime' on pyboard 1.13.0-95
"""
# MCU: (sysname='pyboard', nodename='pyboard', release='1.13.0', version='v1.13-95-g0fff2e03f on 2020-10-03', machine='PYBv1.1 with STM32F405RG')
# Stubber: 1.3.4 - updated
from typing import Any
def gmtime(*args) -> Any:
pass
def localtime(*args) -> Any:
pass
def mktime(*args) -> Any:
pass
def sleep(*args) -> Any:
pass
def sleep_ms(*args) -> Any:
pass
def sleep_us(*args) -> Any:
pass
def ticks_add(*args) -> Any:
pass
def ticks_cpu(*args) -> Any:
pass
def ticks_diff(*args) -> Any:
pass
def ticks_ms(*args) -> Any:
pass
def ticks_us(*args) -> Any:
pass
def time(*args) -> Any:
pass
def time_ns(*args) -> Any:
pass
| 12.559322 | 145 | 0.60054 | 113 | 741 | 3.867257 | 0.39823 | 0.208238 | 0.327231 | 0.384439 | 0.448513 | 0 | 0 | 0 | 0 | 0 | 0 | 0.064799 | 0.22942 | 741 | 58 | 146 | 12.775862 | 0.700525 | 0.278003 | 0 | 0.481481 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.481481 | true | 0.481481 | 0.037037 | 0 | 0.518519 | 0 | 0 | 0 | 0 | null | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 0 | 0 | 1 | 0 | 0 | 7 |
a4aaaaece86d0c63ed9292f64859cb26453869ba | 3,806 | py | Python | utility/refined_events/dialer_record.py | EfficientAI/efficient_cv | e308f229e4d99da86ad56f87f3a78b2c81f27ca5 | [
"MIT"
] | null | null | null | utility/refined_events/dialer_record.py | EfficientAI/efficient_cv | e308f229e4d99da86ad56f87f3a78b2c81f27ca5 | [
"MIT"
] | null | null | null | utility/refined_events/dialer_record.py | EfficientAI/efficient_cv | e308f229e4d99da86ad56f87f3a78b2c81f27ca5 | [
"MIT"
] | null | null | null | from com.android.monkeyrunner import MonkeyRunner
from com.android.monkeyrunner import MonkeyDevice
print('Connecting to device...')
device = MonkeyRunner.waitForConnection()
print('Connected to device')
# Reproduce action log from here
print('Start to reproduce action log')
device.touch(536, 1708, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(536, 1708, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(1.0)
device.touch(148, 312, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(148, 312, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(1.0)
device.touch(950, 1750, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(118, 608, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(1.0)
device.touch(199, 1064, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(199, 1064, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(543, 1060, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(543, 1060, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(867, 1096, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(867, 1096, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(182, 1260, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(182, 1260, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(560, 1244, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(560, 1244, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(897, 1232, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(897, 1232, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(219, 1412, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(219, 1412, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(536, 1432, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(536, 1432, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(914, 1408, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(914, 1408, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(0.2)
device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(992, 872, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(1.0)
device.touch(816, 228, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(816, 228, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(1.0)
device.touch(381, 656, MonkeyDevice.DOWN_AND_UP)
print('Executing : device.touch(381, 656, MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(1.0)
device.press("KEYCODE_HOME", MonkeyDevice.DOWN_AND_UP)
print('Executing : device.press("KEYCODE_HOME", MonkeyDevice.DOWN_AND_UP)')
MonkeyRunner.sleep(1.0)
print('Finish to reproduce action log') | 35.90566 | 75 | 0.788229 | 568 | 3,806 | 5.109155 | 0.107394 | 0.264645 | 0.314266 | 0.347347 | 0.929704 | 0.90765 | 0.90765 | 0.902826 | 0.616127 | 0.566506 | 0 | 0.097623 | 0.071466 | 3,806 | 106 | 76 | 35.90566 | 0.723543 | 0.007882 | 0 | 0.531646 | 0 | 0 | 0.415629 | 0.166358 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.025316 | 0 | 0.025316 | 0.35443 | 0 | 0 | 0 | null | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
101f35eaf0e79f3f32b7e5bb7d3da1592de4065c | 10,439 | py | Python | dialogue-engine/test/programytest/parser/template/graph_tests/test_set.py | cotobadesign/cotoba-agent-oss | 3833d56e79dcd7529c3e8b3a3a8a782d513d9b12 | [
"MIT"
] | 104 | 2020-03-30T09:40:00.000Z | 2022-03-06T22:34:25.000Z | dialogue-engine/test/programytest/parser/template/graph_tests/test_set.py | cotobadesign/cotoba-agent-oss | 3833d56e79dcd7529c3e8b3a3a8a782d513d9b12 | [
"MIT"
] | 25 | 2020-06-12T01:36:35.000Z | 2022-02-19T07:30:44.000Z | dialogue-engine/test/programytest/parser/template/graph_tests/test_set.py | cotobadesign/cotoba-agent-oss | 3833d56e79dcd7529c3e8b3a3a8a782d513d9b12 | [
"MIT"
] | 10 | 2020-04-02T23:43:56.000Z | 2021-05-14T13:47:01.000Z | """
Copyright (c) 2020 COTOBA DESIGN, Inc.
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated
documentation files (the "Software"), to deal in the Software without restriction, including without limitation
the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software,
and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO
THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
"""
import xml.etree.ElementTree as ET
from programy.parser.template.nodes.base import TemplateNode
from programy.parser.template.nodes.set import TemplateSetNode
from programy.parser.exceptions import ParserException
from programytest.parser.template.graph_tests.graph_test_client import TemplateGraphTestClient
class TemplateGraphSetTests(TemplateGraphTestClient):
def test_set_template_typename_as_attrib(self):
template = ET.fromstring("""
<template>
<set name="somepred">Value1</set>
</template>
""")
ast = self._graph.parse_template_expression(template)
self.assertIsNotNone(ast)
self.assertIsInstance(ast, TemplateNode)
self.assertIsNotNone(ast.children)
self.assertEqual(len(ast.children), 1)
set_node = ast.children[0]
self.assertIsNotNone(set_node)
self.assertIsInstance(set_node, TemplateSetNode)
self.assertIsNotNone(set_node.name)
self.assertIsInstance(set_node.name, TemplateNode)
self.assertEqual(set_node.name.resolve(self._client_context), "somepred")
self.assertEqual(set_node.property_type, "name")
self.assertEqual(len(set_node.children), 1)
self.assertEqual(set_node.children[0].resolve(self._client_context), "Value1")
def test_set_template_multi_word_typename_as_attrib(self):
template = ET.fromstring("""
<template>
<set name="somepred other">Value1</set>
</template>
""")
ast = self._graph.parse_template_expression(template)
self.assertIsNotNone(ast)
self.assertIsInstance(ast, TemplateNode)
self.assertIsNotNone(ast.children)
self.assertEqual(len(ast.children), 1)
set_node = ast.children[0]
self.assertIsNotNone(set_node)
self.assertIsInstance(set_node, TemplateSetNode)
self.assertIsNotNone(set_node.name)
self.assertIsInstance(set_node.name, TemplateNode)
self.assertEqual(set_node.name.resolve(self._client_context), "somepred other")
self.assertEqual(set_node.property_type, "name")
self.assertEqual(len(set_node.children), 1)
self.assertEqual(set_node.children[0].resolve(self._client_context), "Value1")
def test_set_template_typename_nested(self):
template = ET.fromstring("""
<template>
Some text here
<set name="somepred">Value1</set>
Some text there
</template>
""")
ast = self._graph.parse_template_expression(template)
self.assertIsNotNone(ast)
self.assertIsInstance(ast, TemplateNode)
self.assertIsNotNone(ast.children)
self.assertEqual(len(ast.children), 3)
set_node = ast.children[1]
self.assertIsNotNone(set_node)
self.assertIsInstance(set_node, TemplateSetNode)
self.assertIsNotNone(set_node.name)
self.assertIsInstance(set_node.name, TemplateNode)
self.assertEqual(set_node.name.resolve(self._client_context), "somepred")
self.assertEqual(set_node.property_type, "name")
self.assertEqual(len(set_node.children), 1)
self.assertEqual(set_node.children[0].resolve(self._client_context), "Value1")
def test_set_template_typedata_as_attrib(self):
template = ET.fromstring("""
<template>
<set data="somedata">Value2</set>
</template>
""")
ast = self._graph.parse_template_expression(template)
self.assertIsNotNone(ast)
self.assertIsInstance(ast, TemplateNode)
self.assertIsNotNone(ast.children)
self.assertEqual(len(ast.children), 1)
set_node = ast.children[0]
self.assertIsNotNone(set_node)
self.assertIsInstance(set_node, TemplateSetNode)
self.assertIsNotNone(set_node.name)
self.assertIsInstance(set_node.name, TemplateNode)
self.assertEqual(set_node.name.resolve(self._client_context), "somedata")
self.assertEqual(set_node.property_type, "data")
self.assertEqual(len(set_node.children), 1)
self.assertEqual(set_node.children[0].resolve(self._client_context), "Value2")
def test_set_template_typevar_as_attrib(self):
template = ET.fromstring("""
<template>
<set var="somevar">Value3</set>
</template>
""")
ast = self._graph.parse_template_expression(template)
self.assertIsNotNone(ast)
self.assertIsInstance(ast, TemplateNode)
self.assertIsNotNone(ast.children)
self.assertEqual(len(ast.children), 1)
set_node = ast.children[0]
self.assertIsNotNone(set_node)
self.assertIsInstance(set_node, TemplateSetNode)
self.assertIsNotNone(set_node.name)
self.assertIsInstance(set_node.name, TemplateNode)
self.assertEqual(set_node.name.resolve(self._client_context), "somevar")
self.assertEqual(set_node.property_type, "var")
self.assertEqual(len(set_node.children), 1)
self.assertEqual(set_node.children[0].resolve(self._client_context), "Value3")
def test_set_template_typename_as_child(self):
template = ET.fromstring("""
<template>
<set><name>somepred</name>Value4</set>
</template>
""")
ast = self._graph.parse_template_expression(template)
self.assertIsNotNone(ast)
self.assertIsInstance(ast, TemplateNode)
self.assertIsNotNone(ast.children)
self.assertEqual(len(ast.children), 1)
set_node = ast.children[0]
self.assertIsNotNone(set_node)
self.assertIsInstance(set_node, TemplateSetNode)
self.assertIsNotNone(set_node.name)
self.assertIsInstance(set_node.name, TemplateNode)
self.assertEqual(set_node.name.resolve(self._client_context), "somepred")
self.assertEqual(set_node.property_type, "name")
self.assertEqual(len(set_node.children), 1)
self.assertEqual(set_node.children[0].resolve(self._client_context), "Value4")
def test_set_template_typedata_as_child(self):
template = ET.fromstring("""
<template>
<set><data>somedata</data>Value5</set>
</template>
""")
ast = self._graph.parse_template_expression(template)
self.assertIsNotNone(ast)
self.assertIsInstance(ast, TemplateNode)
self.assertIsNotNone(ast.children)
self.assertEqual(len(ast.children), 1)
set_node = ast.children[0]
self.assertIsNotNone(set_node)
self.assertIsInstance(set_node, TemplateSetNode)
self.assertIsNotNone(set_node.name)
self.assertIsInstance(set_node.name, TemplateNode)
self.assertEqual(set_node.name.resolve(self._client_context), "somedata")
self.assertEqual(set_node.property_type, "data")
self.assertEqual(len(set_node.children), 1)
self.assertEqual(set_node.children[0].resolve(self._client_context), "Value5")
def test_set_template_typevar_as_child(self):
template = ET.fromstring("""
<template>
<set><var>somevar</var>Value6</set>
</template>
""")
ast = self._graph.parse_template_expression(template)
self.assertIsNotNone(ast)
self.assertIsInstance(ast, TemplateNode)
self.assertIsNotNone(ast.children)
self.assertEqual(len(ast.children), 1)
set_node = ast.children[0]
self.assertIsNotNone(set_node)
self.assertIsInstance(set_node, TemplateSetNode)
self.assertIsNotNone(set_node.name)
self.assertIsInstance(set_node.name, TemplateNode)
self.assertEqual(set_node.name.resolve(self._client_context), "somevar")
self.assertEqual(set_node.property_type, "var")
self.assertEqual(len(set_node.children), 1)
self.assertEqual(set_node.children[0].resolve(self._client_context), "Value6")
def test_set_type_name_and_var(self):
template = ET.fromstring("""
<template>
<set name="somepred" var="somevar">Value1</set>
</template>
""")
with self.assertRaises(ParserException):
self._graph.parse_template_expression(template)
def test_set_type_name_and_data(self):
template = ET.fromstring("""
<template>
<set name="somepred" data="somedata">Value1</set>
</template>
""")
with self.assertRaises(ParserException):
self._graph.parse_template_expression(template)
def test_set_type_data_and_var(self):
template = ET.fromstring("""
<template>
<set data="somedata" var="somevar">Value1</set>
</template>
""")
with self.assertRaises(ParserException):
self._graph.parse_template_expression(template)
def test_set_other(self):
template = ET.fromstring("""
<template>
<set>Value1</set>
</template>
""")
with self.assertRaises(ParserException):
self._graph.parse_template_expression(template)
| 42.092742 | 126 | 0.666635 | 1,152 | 10,439 | 5.859375 | 0.141493 | 0.074667 | 0.039111 | 0.078222 | 0.810667 | 0.792 | 0.764148 | 0.764148 | 0.701778 | 0.701778 | 0 | 0.006978 | 0.231248 | 10,439 | 247 | 127 | 42.263158 | 0.834143 | 0.101734 | 0 | 0.81 | 0 | 0 | 0.174725 | 0.032127 | 0 | 0 | 0 | 0 | 0.5 | 1 | 0.06 | false | 0 | 0.025 | 0 | 0.09 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 |
103f2521cefdc18868ad041de3f7cb2d282e2f30 | 1,952 | py | Python | heltour/tournament/migrations/0075_auto_20160818_0456.py | lenguyenthanh/heltour | 13018b1905539de0b273370a76f6aa1d1ebbb01a | [
"MIT"
] | null | null | null | heltour/tournament/migrations/0075_auto_20160818_0456.py | lenguyenthanh/heltour | 13018b1905539de0b273370a76f6aa1d1ebbb01a | [
"MIT"
] | null | null | null | heltour/tournament/migrations/0075_auto_20160818_0456.py | lenguyenthanh/heltour | 13018b1905539de0b273370a76f6aa1d1ebbb01a | [
"MIT"
] | null | null | null | # -*- coding: utf-8 -*-
# Generated by Django 1.9.7 on 2016-08-18 04:56
from __future__ import unicode_literals
from django.db import migrations
import heltour.tournament.models
class Migration(migrations.Migration):
dependencies = [
('tournament', '0074_auto_20160817_2247'),
]
operations = [
migrations.AlterField(
model_name='loneplayerscore',
name='late_join_points',
field=heltour.tournament.models.ScoreField(default=0),
),
migrations.AlterField(
model_name='loneplayerscore',
name='points',
field=heltour.tournament.models.ScoreField(default=0),
),
migrations.AlterField(
model_name='loneplayerscore',
name='tiebreak1',
field=heltour.tournament.models.ScoreField(default=0),
),
migrations.AlterField(
model_name='loneplayerscore',
name='tiebreak2',
field=heltour.tournament.models.ScoreField(default=0),
),
migrations.AlterField(
model_name='loneplayerscore',
name='tiebreak3',
field=heltour.tournament.models.ScoreField(default=0),
),
migrations.AlterField(
model_name='loneplayerscore',
name='tiebreak4',
field=heltour.tournament.models.ScoreField(default=0),
),
migrations.AlterField(
model_name='teampairing',
name='black_points',
field=heltour.tournament.models.ScoreField(default=0),
),
migrations.AlterField(
model_name='teampairing',
name='white_points',
field=heltour.tournament.models.ScoreField(default=0),
),
migrations.AlterField(
model_name='teamscore',
name='game_points',
field=heltour.tournament.models.ScoreField(default=0),
),
]
| 31.483871 | 66 | 0.59375 | 171 | 1,952 | 6.649123 | 0.309942 | 0.149516 | 0.202287 | 0.229551 | 0.746702 | 0.746702 | 0.704485 | 0.704485 | 0.658751 | 0.658751 | 0 | 0.032727 | 0.295594 | 1,952 | 61 | 67 | 32 | 0.794182 | 0.034324 | 0 | 0.648148 | 1 | 0 | 0.131243 | 0.012221 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.055556 | 0 | 0.111111 | 0 | 0 | 0 | 0 | null | 0 | 1 | 1 | 0 | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
10487f6ac5c71017ff49947a3dbd1c852de25b8e | 59 | py | Python | src/ckanext-uebpackage/ckanext/uebpackage/celery_import.py | CI-WATER/portal | c61660c8389c7af82517cbd0154bc83f9737c4d1 | [
"BSD-3-Clause"
] | 1 | 2017-12-22T04:53:06.000Z | 2017-12-22T04:53:06.000Z | src/ckanext-uebpackage/ckanext/uebpackage/celery_import.py | CI-WATER/portal | c61660c8389c7af82517cbd0154bc83f9737c4d1 | [
"BSD-3-Clause"
] | 4 | 2015-04-16T19:41:17.000Z | 2020-10-25T18:02:14.000Z | src/ckanext-uebpackage/ckanext/uebpackage/celery_import.py | CI-WATER/portal | c61660c8389c7af82517cbd0154bc83f9737c4d1 | [
"BSD-3-Clause"
] | null | null | null | def task_imports():
return ['ckanext.uebpackage.tasks'] | 29.5 | 39 | 0.728814 | 7 | 59 | 6 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.118644 | 59 | 2 | 39 | 29.5 | 0.807692 | 0 | 0 | 0 | 0 | 0 | 0.4 | 0.4 | 0 | 0 | 0 | 0 | 0 | 1 | 0.5 | true | 0 | 0.5 | 0.5 | 1.5 | 0 | 1 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 1 | 1 | 1 | 0 | 0 | 7 |
1085f313ac17a9f85f8d984b98c9da2c9b3da552 | 113 | py | Python | pfdicom_tagSub/__init__.py | reddigari/pfdicom_tagSub | 0eb8f7be164f6f59db35b18ab94b25b887b8163e | [
"Apache-2.0"
] | null | null | null | pfdicom_tagSub/__init__.py | reddigari/pfdicom_tagSub | 0eb8f7be164f6f59db35b18ab94b25b887b8163e | [
"Apache-2.0"
] | 1 | 2021-04-08T18:52:04.000Z | 2021-04-08T18:52:04.000Z | pfdicom_tagSub/__init__.py | reddigari/pfdicom_tagSub | 0eb8f7be164f6f59db35b18ab94b25b887b8163e | [
"Apache-2.0"
] | 1 | 2020-12-03T18:09:37.000Z | 2020-12-03T18:09:37.000Z | try:
from .pfdicom_tagSub import pfdicom_tagSub
except:
from pfdicom_tagSub import pfdicom_tagSub
| 22.6 | 49 | 0.752212 | 14 | 113 | 5.785714 | 0.428571 | 0.641975 | 0.419753 | 0.567901 | 0.888889 | 0.888889 | 0 | 0 | 0 | 0 | 0 | 0 | 0.221239 | 113 | 4 | 50 | 28.25 | 0.920455 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 0.5 | 0 | 0.5 | 0 | 1 | 0 | 0 | null | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 9 |
52cf61493d0499e13fff8039d5a57667cca9b210 | 47,656 | py | Python | src/genie/libs/parser/iosxr/tests/test_show_igmp.py | nujo/genieparser | 083b01efc46afc32abe1a1858729578beab50cd3 | [
"Apache-2.0"
] | 2 | 2021-01-27T03:37:39.000Z | 2021-01-27T03:40:50.000Z | src/genie/libs/parser/iosxr/tests/test_show_igmp.py | nujo/genieparser | 083b01efc46afc32abe1a1858729578beab50cd3 | [
"Apache-2.0"
] | 1 | 2020-08-01T00:23:31.000Z | 2020-08-01T00:40:05.000Z | src/genie/libs/parser/iosxr/tests/test_show_igmp.py | nujo/genieparser | 083b01efc46afc32abe1a1858729578beab50cd3 | [
"Apache-2.0"
] | null | null | null |
import re
import unittest
from unittest.mock import Mock
from pyats.topology import Device
from genie.metaparser.util.exceptions import SchemaEmptyParserError, \
SchemaMissingKeyError
from genie.libs.parser.iosxr.show_igmp import ShowIgmpInterface, ShowIgmpSummary, ShowIgmpGroupsDetail
#############################################################################
# unitest For Show IGMP Interface
#############################################################################
class TestShowIgmpInterface(unittest.TestCase):
device = Device(name='aDevice')
device0 = Device(name='bDevice')
empty_output = {'execute.return_value': ''}
golden_parsed_output = {
'vrf': {
'default': {
'interfaces': {
'GigabitEthernet0/0/0/0.110': {
'interface_status': 'up',
'igmp_activity': {
'joins': 7,
'leaves': 1
},
'igmp_max_query_response_time': 10,
'igmp_querier_timeout': 125,
'igmp_query_interval': 60,
'last_member_query_response_interval': 1,
'igmp_querying_router': '10.12.110.1',
'igmp_state': 'enabled',
'time_elapsed_since_last_query_sent': '02:42:58',
'time_elapsed_since_last_report_received': '00:00:31',
'time_elapsed_since_router_enabled': '02:46:41',
'igmp_version': 3,
'ip_address': '10.12.110.2/24',
'line_protocol': 'up',
'oper_status': 'up'
},
'GigabitEthernet0/0/0/0.115': {
'interface_status': 'up',
'igmp_state': 'disabled',
'ip_address': '10.12.115.2/24',
'line_protocol': 'up',
'oper_status': 'up'
},
'GigabitEthernet0/0/0/0.120': {
'interface_status': 'up',
'igmp_state': 'disabled',
'ip_address': '10.12.120.2/24',
'line_protocol': 'up',
'oper_status': 'up'
},
'GigabitEthernet0/0/0/0.90':{
'interface_status': 'up',
'igmp_state': 'disabled',
'ip_address': '10.12.90.2/24',
'line_protocol': 'up',
'oper_status': 'up'
},
'GigabitEthernet0/0/0/1.110': {
'interface_status': 'up',
'igmp_activity': {
'joins': 5,
'leaves': 0
},
'igmp_max_query_response_time': 10,
'igmp_querier_timeout': 125,
'igmp_query_interval': 60,
'last_member_query_response_interval': 1,
'igmp_querying_router': '10.23.110.2',
'igmp_querying_router_info': 'this system',
'igmp_state': 'enabled',
'time_elapsed_since_last_query_sent': '00:00:55',
'time_elapsed_since_last_report_received': '00:00:55',
'time_elapsed_since_router_enabled': '02:46:41',
'igmp_version': 3,
'ip_address': '10.23.110.2/24',
'line_protocol': 'up',
'oper_status': 'up'
},
'GigabitEthernet0/0/0/1.115': {
'interface_status': 'up',
'igmp_state': 'disabled',
'ip_address': '10.23.115.2/24',
'line_protocol': 'up',
'oper_status': 'up'
},
'GigabitEthernet0/0/0/1.120': {
'interface_status': 'up',
'igmp_state': 'disabled',
'ip_address': '10.23.120.2/24',
'line_protocol': 'up',
'oper_status': 'up'
},
'GigabitEthernet0/0/0/1.90': {
'interface_status': 'up',
'igmp_state': 'disabled',
'ip_address': '10.23.90.2/24',
'line_protocol': 'up',
'oper_status': 'up'
},
'Loopback0': {
'interface_status': 'up',
'igmp_activity': {
'joins': 6,
'leaves': 0
},
'igmp_max_query_response_time': 10,
'igmp_querier_timeout': 125,
'igmp_query_interval': 60,
'last_member_query_response_interval': 1,
'igmp_querying_router': '10.16.2.2',
'igmp_querying_router_info': 'this system',
'igmp_state': 'enabled',
'time_elapsed_since_last_query_sent': '00:00:53',
'time_elapsed_since_last_report_received': '00:00:51',
'time_elapsed_since_router_enabled': '02:46:41',
'igmp_version': 3,
'ip_address': '10.16.2.2/32',
'line_protocol': 'up',
'oper_status': 'up'
}
}
}
}
}
golden_output = {'execute.return_value': '''
R2_xr# show igmp interface
Loopback0 is up, line protocol is up
Internet address is 10.16.2.2/32
IGMP is enabled on interface
Current IGMP version is 3
IGMP query interval is 60 seconds
IGMP querier timeout is 125 seconds
IGMP max query response time is 10 seconds
Last member query response interval is 1 seconds
IGMP activity: 6 joins, 0 leaves
IGMP querying router is 10.16.2.2 (this system)
Time elapsed since last query sent 00:00:53
Time elapsed since IGMP router enabled 02:46:41
Time elapsed since last report received 00:00:51
GigabitEthernet0/0/0/0.90 is up, line protocol is up
Internet address is 10.12.90.2/24
IGMP is disabled on interface
GigabitEthernet0/0/0/1.90 is up, line protocol is up
Internet address is 10.23.90.2/24
IGMP is disabled on interface
GigabitEthernet0/0/0/0.110 is up, line protocol is up
Internet address is 10.12.110.2/24
IGMP is enabled on interface
Current IGMP version is 3
IGMP query interval is 60 seconds
IGMP querier timeout is 125 seconds
IGMP max query response time is 10 seconds
Last member query response interval is 1 seconds
IGMP activity: 7 joins, 1 leaves
IGMP querying router is 10.12.110.1
Time elapsed since last query sent 02:42:58
Time elapsed since IGMP router enabled 02:46:41
Time elapsed since last report received 00:00:31
GigabitEthernet0/0/0/0.115 is up, line protocol is up
Internet address is 10.12.115.2/24
IGMP is disabled on interface
GigabitEthernet0/0/0/0.120 is up, line protocol is up
Internet address is 10.12.120.2/24
IGMP is disabled on interface
GigabitEthernet0/0/0/1.110 is up, line protocol is up
Internet address is 10.23.110.2/24
IGMP is enabled on interface
Current IGMP version is 3
IGMP query interval is 60 seconds
IGMP querier timeout is 125 seconds
IGMP max query response time is 10 seconds
Last member query response interval is 1 seconds
IGMP activity: 5 joins, 0 leaves
IGMP querying router is 10.23.110.2 (this system)
Time elapsed since last query sent 00:00:55
Time elapsed since IGMP router enabled 02:46:41
Time elapsed since last report received 00:00:55
GigabitEthernet0/0/0/1.115 is up, line protocol is up
Internet address is 10.23.115.2/24
IGMP is disabled on interface
GigabitEthernet0/0/0/1.120 is up, line protocol is up
Internet address is 10.23.120.2/24
IGMP is disabled on interface
'''}
golden_parsed_interface_output = {
'vrf': {
'default': {
'interfaces': {
'GigabitEthernet0/0/0/1.115': {
'interface_status': 'up',
'igmp_state': 'disabled',
'ip_address': '10.23.115.2/24',
'line_protocol': 'up',
'oper_status': 'up'
}
}
}
}
}
golden_parsed_interface_output1 = {
'vrf': {
'VRF1': {
'interfaces': {
'Loopback0': {
'interface_status': 'up',
'igmp_activity': {
'joins': 6,
'leaves': 0
},
'igmp_max_query_response_time': 10,
'igmp_querier_timeout': 125,
'igmp_query_interval': 60,
'last_member_query_response_interval': 1,
'igmp_querying_router': '10.16.2.2',
'igmp_querying_router_info': 'this system',
'igmp_state': 'enabled',
'time_elapsed_since_last_query_sent': '00:00:53',
'time_elapsed_since_last_report_received': '00:00:51',
'time_elapsed_since_router_enabled': '02:46:41',
'igmp_version': 3,
'ip_address': '10.16.2.2/32',
'line_protocol': 'up',
'oper_status': 'up'
}
}
}
}
}
golden_interface_output={'execute.return_value':'''
R2_xr# show igmp interface GigabitEthernet0/0/0/1.115
GigabitEthernet0/0/0/1.115 is up, line protocol is up
Internet address is 10.23.115.2/24
IGMP is disabled on interface
'''}
golden_interface_output1={'execute.return_value':'''
R2_xr# show igmp vrf VRF1 interface
Loopback0 is up, line protocol is up
Internet address is 10.16.2.2/32
IGMP is enabled on interface
Current IGMP version is 3
IGMP query interval is 60 seconds
IGMP querier timeout is 125 seconds
IGMP max query response time is 10 seconds
Last member query response interval is 1 seconds
IGMP activity: 6 joins, 0 leaves
IGMP querying router is 10.16.2.2 (this system)
Time elapsed since last query sent 00:00:53
Time elapsed since IGMP router enabled 02:46:41
Time elapsed since last report received 00:00:51
'''}
golden_parsed_interface_output2 = {
'vrf': {
'VRF1': {
'interfaces': {
'Loopback0': {
'interface_status': 'up',
'igmp_activity': {
'joins': 6,
'leaves': 0
},
'igmp_max_query_response_time': 10,
'igmp_querier_timeout': 125,
'igmp_query_interval': 60,
'last_member_query_response_interval': 1,
'igmp_querying_router': '10.16.2.2',
'igmp_querying_router_info': 'this system',
'igmp_state': 'enabled',
'time_elapsed_since_last_query_sent': '00:00:53',
'time_elapsed_since_last_report_received': '00:00:51',
'time_elapsed_since_router_enabled': '02:46:41',
'igmp_version': 3,
'ip_address': '10.16.2.2/32',
'line_protocol': 'up',
'oper_status': 'up'
}
}
}
}
}
golden_interface_output2={'execute.return_value':'''
R2_xr# show igmp vrf VRF1 interface Loopback0
Loopback0 is up, line protocol is up
Internet address is 10.16.2.2/32
IGMP is enabled on interface
Current IGMP version is 3
IGMP query interval is 60 seconds
IGMP querier timeout is 125 seconds
IGMP max query response time is 10 seconds
Last member query response interval is 1 seconds
IGMP activity: 6 joins, 0 leaves
IGMP querying router is 10.16.2.2 (this system)
Time elapsed since last query sent 00:00:53
Time elapsed since IGMP router enabled 02:46:41
Time elapsed since last report received 00:00:51
'''}
def test_empty(self):
self.device1 = Mock(**self.empty_output)
interface_detail_obj = ShowIgmpInterface(device=self.device1)
with self.assertRaises(SchemaEmptyParserError):
parsed_output = interface_detail_obj.parse()
def test_golden(self):
self.device = Mock(**self.golden_output)
interface_detail_obj = ShowIgmpInterface(device=self.device)
parsed_output = interface_detail_obj.parse()
self.maxDiff = None
self.assertEqual(parsed_output,self.golden_parsed_output)
def test_golden_custom_interface(self):
self.device = Mock(**self.golden_interface_output)
interface_detail_obj = ShowIgmpInterface(device=self.device)
parsed_output = interface_detail_obj.parse(interface='GigabitEthernet0/0/0/1.115')
self.maxDiff = None
self.assertEqual(parsed_output, self.golden_parsed_interface_output)
def test_golden_custom_vrf(self):
self.device = Mock(**self.golden_interface_output1)
interface_detail_obj = ShowIgmpInterface(device=self.device)
parsed_output = interface_detail_obj.parse(vrf='VRF1')
self.maxDiff = None
self.assertEqual(parsed_output, self.golden_parsed_interface_output1)
def test_golden_custom_vrf_interface(self):
self.device = Mock(**self.golden_interface_output2)
interface_detail_obj = ShowIgmpInterface(device=self.device)
parsed_output = interface_detail_obj.parse(vrf='VRF1', interface='Loopback0')
self.maxDiff = None
self.assertEqual(parsed_output, self.golden_parsed_interface_output2)
#############################################################################
# unitest For Show IGMP Summary
#############################################################################
class test_show_igmp_summary(unittest.TestCase):
device = Device(name='aDevice')
device0 = Device(name='bDevice')
empty_output = {'execute.return_value': ''}
golden_parsed_output = {
'vrf': {
'default': {
'disabled_interfaces': 6,
'enabled_interfaces': 3,
'no_of_group_x_interface': 16,
'interfaces': {
'Loopback0': {
'max_groups': 25000,
'number_groups': 6
},
'GigabitEthernet0/0/0/0.90': {
'max_groups': 25000,
'number_groups': 1
},
'GigabitEthernet0/0/0/1.90': {
'max_groups': 25000,
'number_groups': 1
},
'GigabitEthernet0/0/0/0.110': {
'max_groups': 25000,
'number_groups': 6
},
'GigabitEthernet0/0/0/0.115': {
'max_groups': 25000,
'number_groups': 4
},
'GigabitEthernet0/0/0/0.120': {
'max_groups': 25000,
'number_groups': 1
},
'GigabitEthernet0/0/0/1.110': {
'max_groups': 25000,
'number_groups': 5
},
'GigabitEthernet0/0/0/1.115': {
'max_groups': 25000,
'number_groups': 0
},
'GigabitEthernet0/0/0/1.120': {
'max_groups': 25000,
'number_groups': 1
}
},
'mte_tuple_count': 0,
'maximum_number_of_groups_for_vrf': 50000,
'robustness_value': 2,
'supported_interfaces': 9,
'unsupported_interfaces': 0,
}
}
}
golden_output = {'execute.return_value': '''
R2_xr#show igmp summary
Robustness Value 2
No. of Group x Interfaces 16
Maximum number of Groups for this VRF 50000
Supported Interfaces : 9
Unsupported Interfaces : 0
Enabled Interfaces : 3
Disabled Interfaces : 6
MTE tuple count : 0
Interface Number Max #
Groups Groups
Loopback0 6 25000
GigabitEthernet0/0/0/0.90 1 25000
GigabitEthernet0/0/0/1.90 1 25000
GigabitEthernet0/0/0/0.110 6 25000
GigabitEthernet0/0/0/0.115 4 25000
GigabitEthernet0/0/0/0.120 1 25000
GigabitEthernet0/0/0/1.110 5 25000
GigabitEthernet0/0/0/1.115 0 25000
GigabitEthernet0/0/0/1.120 1 25000
'''}
golden_parsed_summary_output = {
'vrf': {
'VRF1': {
'disabled_interfaces': 6,
'enabled_interfaces': 3,
'no_of_group_x_interface': 15,
'interfaces': {
'Loopback300': {
'max_groups': 25000,
'number_groups': 4
},
'GigabitEthernet0/0/0/0.390': {
'max_groups': 25000,
'number_groups': 1
},
'GigabitEthernet0/0/0/0.410': {
'max_groups': 25000,
'number_groups': 7
},
'GigabitEthernet0/0/0/0.415': {
'max_groups': 25000,
'number_groups': 4
},
'GigabitEthernet0/0/0/0.420': {
'max_groups': 25000,
'number_groups': 1
},
'GigabitEthernet0/0/0/1.390': {
'max_groups': 25000,
'number_groups': 1
},
'GigabitEthernet0/0/0/1.410': {
'max_groups': 25000,
'number_groups': 5
},
'GigabitEthernet0/0/0/1.415': {
'max_groups': 25000,
'number_groups': 0
},
'GigabitEthernet0/0/0/1.420': {
'max_groups': 25000,
'number_groups': 1
}
},
'mte_tuple_count': 0,
'maximum_number_of_groups_for_vrf': 50000,
'robustness_value': 2,
'supported_interfaces': 9,
'unsupported_interfaces': 0,
}
}
}
golden_summary_output={'execute.return_value':'''
R2_xr#show igmp vrf VRF1 summary
Robustness Value 2
No. of Group x Interfaces 15
Maximum number of Groups for this VRF 50000
Supported Interfaces : 9
Unsupported Interfaces : 0
Enabled Interfaces : 3
Disabled Interfaces : 6
MTE tuple count : 0
Interface Number Max #
Groups Groups
Loopback300 4 25000
GigabitEthernet0/0/0/0.390 1 25000
GigabitEthernet0/0/0/0.410 7 25000
GigabitEthernet0/0/0/0.415 4 25000
GigabitEthernet0/0/0/0.420 1 25000
GigabitEthernet0/0/0/1.390 1 25000
GigabitEthernet0/0/0/1.410 5 25000
GigabitEthernet0/0/0/1.415 0 25000
GigabitEthernet0/0/0/1.420 1 25000
'''}
def test_empty(self):
self.device1 = Mock(**self.empty_output)
summary_detail_obj = ShowIgmpSummary(device=self.device1)
with self.assertRaises(SchemaEmptyParserError):
parsed_output = summary_detail_obj.parse()
def test_golden(self):
self.device = Mock(**self.golden_output)
summary_detail_obj = ShowIgmpSummary(device=self.device)
parsed_output = summary_detail_obj.parse()
self.maxDiff = None
self.assertEqual(parsed_output,self.golden_parsed_output)
def test_golden_custom(self):
self.device = Mock(**self.golden_summary_output)
summary_detail_obj = ShowIgmpSummary(device=self.device)
parsed_output = summary_detail_obj.parse(vrf='VRF1')
self.maxDiff = None
self.assertEqual(parsed_output, self.golden_parsed_summary_output)
#############################################################################
# unitest For Show IGMP Groups Detail
#############################################################################
class test_show_igmp_groups_detail(unittest.TestCase):
device = Device(name='aDevice')
device0 = Device(name='bDevice')
empty_output = {'execute.return_value': ''}
golden_parsed_output = {
"vrf": {
"default": {
"interfaces": {
"Loopback0": {
"group": {
"224.0.0.2": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:44:55"
},
"224.0.0.9": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "09:47:23"
},
"224.0.0.13": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:44:55"
},
"224.0.0.22": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:44:55"
},
"224.0.1.39": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:19:56"
},
"224.0.1.40": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:44:55"
}
}
},
"GigabitEthernet0/0/0/0.90": {
"group": {
"224.0.0.10": {
"host_mode": "exclude",
"last_reporter": "0.0.0.0",
"router_mode": "INCLUDE",
"router_mode_expires": "None",
"suppress": 0,
"up_time": "01:53:32"
}
}
},
"GigabitEthernet0/0/0/0.110": {
"group": {
"224.0.0.2": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:44:55"
},
"224.0.0.5": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "10:36:57"
},
"224.0.0.6": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "10:36:57"
},
"224.0.0.13": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:44:55"
},
"224.0.0.22": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:44:55"
},
"224.0.1.39": {
"host_mode": "include",
"last_reporter": "10.12.110.1",
"router_mode": "EXCLUDE",
"router_mode_expires": "00:01:41",
"suppress": 0,
"up_time": "02:29:47",
}
}
},
"GigabitEthernet0/0/0/0.120": {
"group": {
"224.0.0.9": {
"host_mode": "exclude",
"last_reporter": "0.0.0.0",
"router_mode": "INCLUDE",
"router_mode_expires": "None",
"suppress": 0,
"up_time": "09:47:23"
}
}
},
"GigabitEthernet0/0/0/1.120": {
"group": {
"224.0.0.9": {
"host_mode": "exclude",
"last_reporter": "0.0.0.0",
"router_mode": "INCLUDE",
"router_mode_expires": "None",
"suppress": 0,
"up_time": "09:47:23"
}
}
}
}
}
}
}
golden_output = {'execute.return_value': '''
R2_xr#show igmp groups detail
Interface: Loopback0
Group: 224.0.0.2
Uptime: 02:44:55
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: Loopback0
Group: 224.0.0.9
Uptime: 09:47:23
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: Loopback0
Group: 224.0.0.13
Uptime: 02:44:55
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: Loopback0
Group: 224.0.0.22
Uptime: 02:44:55
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: Loopback0
Group: 224.0.1.39
Uptime: 02:19:56
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: Loopback0
Group: 224.0.1.40
Uptime: 02:44:55
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.90
Group: 224.0.0.10
Uptime: 01:53:32
Router mode: INCLUDE
Host mode: EXCLUDE
Last reporter: 0.0.0.0
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.110
Group: 224.0.0.2
Uptime: 02:44:55
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.110
Group: 224.0.0.5
Uptime: 10:36:57
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.110
Group: 224.0.0.6
Uptime: 10:36:57
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.110
Group: 224.0.0.13
Uptime: 02:44:55
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.110
Group: 224.0.0.22
Uptime: 02:44:55
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.110
Group: 224.0.1.39
Uptime: 02:29:47
Router mode: EXCLUDE (Expires: 00:01:41)
Host mode: INCLUDE
Last reporter: 10.12.110.1
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.120
Group: 224.0.0.9
Uptime: 09:47:23
Router mode: INCLUDE
Host mode: EXCLUDE
Last reporter: 0.0.0.0
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/1.120
Group: 224.0.0.9
Uptime: 09:47:23
Router mode: INCLUDE
Host mode: EXCLUDE
Last reporter: 0.0.0.0
Suppress: 0
Source list is empty
'''}
golden_parsed_igmp_groups_output = {
"vrf": {
"VRF1": {
"interfaces": {
"Loopback300": {
"group": {
"224.0.0.2": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:43:30"
},
"224.0.0.9": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "09:48:07"
},
"224.0.0.13": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:43:30"
},
"224.0.0.22": {
"host_mode": "exclude",
"last_reporter": "10.16.2.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:43:30"
}
}
},
"GigabitEthernet0/0/0/0.390": {
"group": {
"224.0.0.10": {
"host_mode": "exclude",
"last_reporter": "0.0.0.0",
"router_mode": "INCLUDE",
"router_mode_expires": "None",
"suppress": 0,
"up_time": "01:54:16"
}
}
},
"GigabitEthernet0/0/0/0.410": {
"group": {
"224.0.0.2": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:43:30"
},
"224.0.0.5": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "10:37:41"
},
"224.0.0.6": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "10:37:41"
},
"224.0.0.13": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:43:30"
},
"224.0.0.22": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:43:30"
},
"224.0.1.39": {
"host_mode": "include",
"last_reporter": "10.12.110.1",
"router_mode": "EXCLUDE",
"router_mode_expires": "00:01:21",
"suppress": 0,
"up_time": "02:30:06"
},
"224.0.1.40": {
"host_mode": "exclude",
"last_reporter": "10.12.110.2",
"router_mode": "EXCLUDE",
"router_mode_expires": "never",
"suppress": 0,
"up_time": "02:43:30"
}
}
},
"GigabitEthernet0/0/0/0.420": {
"group": {
"224.0.0.9": {
"host_mode": "exclude",
"last_reporter": "0.0.0.0",
"router_mode": "INCLUDE",
"router_mode_expires": "None",
"suppress": 0,
"up_time": "09:48:07"
}
}
},
"GigabitEthernet0/0/0/1.390": {
"group": {
"224.0.0.10": {
"host_mode": "exclude",
"last_reporter": "0.0.0.0",
"router_mode": "INCLUDE",
"router_mode_expires": "None",
"suppress": 0,
"up_time": "01:54:16"
}
}
},
"GigabitEthernet0/0/0/1.420": {
"group": {
"224.0.0.9": {
"host_mode": "exclude",
"last_reporter": "0.0.0.0",
"router_mode": "INCLUDE",
"router_mode_expires": "None",
"suppress": 0,
"up_time": "09:48:07"
}
}
}
}
}
}
}
golden_igmp_groups_output={'execute.return_value':'''
R2_xr#show igmp vrf VRF1 groups detail
Interface: Loopback300
Group: 224.0.0.2
Uptime: 02:43:30
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: Loopback300
Group: 224.0.0.9
Uptime: 09:48:07
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: Loopback300
Group: 224.0.0.13
Uptime: 02:43:30
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: Loopback300
Group: 224.0.0.22
Uptime: 02:43:30
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.16.2.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.390
Group: 224.0.0.10
Uptime: 01:54:16
Router mode: INCLUDE
Host mode: EXCLUDE
Last reporter: 0.0.0.0
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.410
Group: 224.0.0.2
Uptime: 02:43:30
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.410
Group: 224.0.0.5
Uptime: 10:37:41
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.410
Group: 224.0.0.6
Uptime: 10:37:41
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.410
Group: 224.0.0.13
Uptime: 02:43:30
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.410
Group: 224.0.0.22
Uptime: 02:43:30
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.410
Group: 224.0.1.39
Uptime: 02:30:06
Router mode: EXCLUDE (Expires: 00:01:21)
Host mode: INCLUDE
Last reporter: 10.12.110.1
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.410
Group: 224.0.1.40
Uptime: 02:43:30
Router mode: EXCLUDE (Expires: never)
Host mode: EXCLUDE
Last reporter: 10.12.110.2
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/0.420
Group: 224.0.0.9
Uptime: 09:48:07
Router mode: INCLUDE
Host mode: EXCLUDE
Last reporter: 0.0.0.0
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/1.390
Group: 224.0.0.10
Uptime: 01:54:16
Router mode: INCLUDE
Host mode: EXCLUDE
Last reporter: 0.0.0.0
Suppress: 0
Source list is empty
Interface: GigabitEthernet0/0/0/1.420
Group: 224.0.0.9
Uptime: 09:48:07
Router mode: INCLUDE
Host mode: EXCLUDE
Last reporter: 0.0.0.0
Suppress: 0
Source list is empty
'''}
golden_output1 = {'execute.return_value': '''
R2_xr#show igmp groups detail
Interface: GigabitEthernet0/0/0/2
Group: 232.1.1.1
Uptime: 00:04:55
Router mode: INCLUDE
Host mode: INCLUDE
Last reporter: 192.168.1.42
Group source list:
Source Address Uptime Expires Fwd Flags
192.168.1.18 00:04:55 00:01:28 Yes Remote
'''}
golden_parsed_output1 = {
"vrf": {
"default": {
"interfaces": {
"GigabitEthernet0/0/0/2": {
"group": {
"232.1.1.1": {
"host_mode": "include",
"last_reporter": "192.168.1.42",
"router_mode": "INCLUDE",
"router_mode_expires": "None",
"up_time": "00:04:55",
"source": {
"192.168.1.18": {
"up_time": "00:04:55",
"expire": "00:01:28",
"forward": "Yes",
"flags": "Remote"
}
}
}
}
}
}
}
}
}
def test_empty(self):
self.device1 = Mock(**self.empty_output)
igmp_groups_detail_obj = ShowIgmpGroupsDetail(device=self.device1)
with self.assertRaises(SchemaEmptyParserError):
parsed_output = igmp_groups_detail_obj.parse()
def test_golden(self):
self.device = Mock(**self.golden_output)
igmp_groups_detail_obj = ShowIgmpGroupsDetail(device=self.device)
parsed_output = igmp_groups_detail_obj.parse()
self.maxDiff = None
self.assertEqual(parsed_output,self.golden_parsed_output)
def test_golden_custom(self):
self.device = Mock(**self.golden_igmp_groups_output)
igmp_groups_detail_obj = ShowIgmpGroupsDetail(device=self.device)
parsed_output = igmp_groups_detail_obj.parse(vrf='VRF1')
self.maxDiff = None
self.assertEqual(parsed_output, self.golden_parsed_igmp_groups_output)
def test_golden1(self):
self.device = Mock(**self.golden_output1)
igmp_groups_detail_obj = ShowIgmpGroupsDetail(device=self.device)
parsed_output = igmp_groups_detail_obj.parse()
self.maxDiff = None
self.assertEqual(parsed_output,self.golden_parsed_output1)
if __name__ == '__main__':
unittest.main()
| 41.047373 | 102 | 0.419444 | 4,396 | 47,656 | 4.388763 | 0.045723 | 0.023014 | 0.077437 | 0.05515 | 0.94739 | 0.932359 | 0.896854 | 0.879127 | 0.867465 | 0.833204 | 0 | 0.112432 | 0.477422 | 47,656 | 1,160 | 103 | 41.082759 | 0.662263 | 0.002035 | 0 | 0.729656 | 0 | 0 | 0.486452 | 0.06878 | 0 | 0 | 0 | 0 | 0.01085 | 1 | 0.01085 | false | 0 | 0.005425 | 0 | 0.0434 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
52f9e2c4f204a43900b04f8eb2b268c771a17bf9 | 36 | py | Python | ex1_5_execfile/simple.py | demsheng/QtBoostPythonHelloWorld | fae98d11c009495cc3b060a1980a0463652768da | [
"MIT"
] | null | null | null | ex1_5_execfile/simple.py | demsheng/QtBoostPythonHelloWorld | fae98d11c009495cc3b060a1980a0463652768da | [
"MIT"
] | null | null | null | ex1_5_execfile/simple.py | demsheng/QtBoostPythonHelloWorld | fae98d11c009495cc3b060a1980a0463652768da | [
"MIT"
] | null | null | null | def foo(i = 4):
return i + 2008
| 12 | 19 | 0.527778 | 7 | 36 | 2.714286 | 0.857143 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.208333 | 0.333333 | 36 | 2 | 20 | 18 | 0.583333 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.5 | false | 0 | 0 | 0.5 | 1 | 0 | 1 | 1 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 1 | 1 | 0 | 0 | 7 |
eaad8e76c5419ad245a5e376f9ee44485551f19c | 45,240 | py | Python | tsupy/go.py | DataIntelligenceCrew/tsupy | 363ecf5616ce4c2367b5550e6351548053a71bb6 | [
"MIT"
] | 1 | 2022-03-25T03:19:17.000Z | 2022-03-25T03:19:17.000Z | tsupy/go.py | DataIntelligenceCrew/tsupy | 363ecf5616ce4c2367b5550e6351548053a71bb6 | [
"MIT"
] | null | null | null | tsupy/go.py | DataIntelligenceCrew/tsupy | 363ecf5616ce4c2367b5550e6351548053a71bb6 | [
"MIT"
] | 2 | 2022-03-25T03:19:21.000Z | 2022-03-30T04:02:55.000Z |
# python wrapper for package go within overall package tsubasa
# This is what you import to use the package.
# File is generated by gopy. Do not edit.
# gopy gen -output=tsupy -vm=python3 github.com/js061/tsubasa
#
# the following is required to enable dlopen to open the _go.so file
import os,sys,inspect,collections
try:
import collections.abc as _collections_abc
except ImportError:
_collections_abc = collections
cwd = os.getcwd()
currentdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))
os.chdir(currentdir)
from . import _tsubasa
os.chdir(cwd)
# to use this code in your end-user python file, import it as follows:
# from tsubasa import go
# and then refer to everything using go. prefix
# packages imported by this package listed below:
import collections
try:
import collections.abc as _collections_abc
except ImportError:
_collections_abc = collections
class GoClass(object):
"""GoClass is the base class for all GoPy wrapper classes"""
def __init__(self):
self.handle = 0
# use go.nil for nil pointers
nil = GoClass()
# need to explicitly initialize it
def main():
global nil
nil = GoClass()
main()
def Init():
"""calls the GoPyInit function, which runs the 'main' code string that was passed using -main arg to gopy"""
_tsubasa.GoPyInit()
# ---- Types ---
# Python type for slice []bool
class Slice_bool(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_bool_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_bool.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_bool len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_bool([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_bool_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_bool_len(self.handle)
return Slice_bool(handle=_tsubasa.Slice_bool_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_bool_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_bool_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_bool.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_bool_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_bool_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []byte
class Slice_byte(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_byte_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_byte.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_byte len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_byte([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_byte_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_byte_len(self.handle)
return Slice_byte(handle=_tsubasa.Slice_byte_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_byte_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_byte_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_byte.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_byte_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_byte_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []float32
class Slice_float32(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_float32_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_float32.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_float32 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_float32([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_float32_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_float32_len(self.handle)
return Slice_float32(handle=_tsubasa.Slice_float32_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_float32_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_float32_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_float32.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_float32_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_float32_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []float64
class Slice_float64(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_float64_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_float64.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_float64 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_float64([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_float64_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_float64_len(self.handle)
return Slice_float64(handle=_tsubasa.Slice_float64_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_float64_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_float64_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_float64.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_float64_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_float64_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []int
class Slice_int(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_int_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_int.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_int len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_int([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_int_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_int_len(self.handle)
return Slice_int(handle=_tsubasa.Slice_int_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_int_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_int_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_int.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_int_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_int_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []int16
class Slice_int16(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_int16_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_int16.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_int16 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_int16([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_int16_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_int16_len(self.handle)
return Slice_int16(handle=_tsubasa.Slice_int16_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_int16_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_int16_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_int16.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_int16_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_int16_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []int32
class Slice_int32(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_int32_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_int32.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_int32 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_int32([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_int32_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_int32_len(self.handle)
return Slice_int32(handle=_tsubasa.Slice_int32_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_int32_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_int32_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_int32.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_int32_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_int32_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []int64
class Slice_int64(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_int64_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_int64.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_int64 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_int64([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_int64_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_int64_len(self.handle)
return Slice_int64(handle=_tsubasa.Slice_int64_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_int64_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_int64_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_int64.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_int64_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_int64_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []int8
class Slice_int8(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_int8_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_int8.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_int8 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_int8([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_int8_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_int8_len(self.handle)
return Slice_int8(handle=_tsubasa.Slice_int8_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_int8_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_int8_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_int8.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_int8_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_int8_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []rune
class Slice_rune(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_rune_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_rune.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_rune len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_rune([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_rune_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_rune_len(self.handle)
return Slice_rune(handle=_tsubasa.Slice_rune_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_rune_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_rune_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_rune.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_rune_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_rune_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []string
class Slice_string(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_string_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_string.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_string len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_string([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_string_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_string_len(self.handle)
return Slice_string(handle=_tsubasa.Slice_string_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_string_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_string_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_string.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_string_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_string_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []uint
class Slice_uint(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_uint_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_uint.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_uint len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_uint([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_uint_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_uint_len(self.handle)
return Slice_uint(handle=_tsubasa.Slice_uint_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_uint_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_uint_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_uint.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_uint_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_uint_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []uint16
class Slice_uint16(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_uint16_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_uint16.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_uint16 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_uint16([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_uint16_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_uint16_len(self.handle)
return Slice_uint16(handle=_tsubasa.Slice_uint16_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_uint16_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_uint16_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_uint16.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_uint16_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_uint16_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []uint32
class Slice_uint32(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_uint32_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_uint32.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_uint32 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_uint32([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_uint32_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_uint32_len(self.handle)
return Slice_uint32(handle=_tsubasa.Slice_uint32_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_uint32_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_uint32_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_uint32.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_uint32_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_uint32_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []uint64
class Slice_uint64(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_uint64_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_uint64.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_uint64 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_uint64([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_uint64_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_uint64_len(self.handle)
return Slice_uint64(handle=_tsubasa.Slice_uint64_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_uint64_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_uint64_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_uint64.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_uint64_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_uint64_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# Python type for slice []uint8
class Slice_uint8(GoClass):
""""""
def __init__(self, *args, **kwargs):
"""
handle=A Go-side object is always initialized with an explicit handle=arg
otherwise parameter is a python list that we copy from
"""
self.index = 0
if len(kwargs) == 1 and 'handle' in kwargs:
self.handle = kwargs['handle']
_tsubasa.IncRef(self.handle)
elif len(args) == 1 and isinstance(args[0], GoClass):
self.handle = args[0].handle
_tsubasa.IncRef(self.handle)
else:
self.handle = _tsubasa.Slice_uint8_CTor()
_tsubasa.IncRef(self.handle)
if len(args) > 0:
if not isinstance(args[0], _collections_abc.Iterable):
raise TypeError('Slice_uint8.__init__ takes a sequence as argument')
for elt in args[0]:
self.append(elt)
def __del__(self):
_tsubasa.DecRef(self.handle)
def __str__(self):
s = 'go.Slice_uint8 len: ' + str(len(self)) + ' handle: ' + str(self.handle) + ' ['
if len(self) < 120:
s += ', '.join(map(str, self)) + ']'
return s
def __repr__(self):
return 'go.Slice_uint8([' + ', '.join(map(str, self)) + '])'
def __len__(self):
return _tsubasa.Slice_uint8_len(self.handle)
def __getitem__(self, key):
if isinstance(key, slice):
if key.step == None or key.step == 1:
st = key.start
ed = key.stop
if st == None:
st = 0
if ed == None:
ed = _tsubasa.Slice_uint8_len(self.handle)
return Slice_uint8(handle=_tsubasa.Slice_uint8_subslice(self.handle, st, ed))
return [self[ii] for ii in range(*key.indices(len(self)))]
elif isinstance(key, int):
if key < 0:
key += len(self)
if key < 0 or key >= len(self):
raise IndexError('slice index out of range')
return _tsubasa.Slice_uint8_elem(self.handle, key)
else:
raise TypeError('slice index invalid type')
def __setitem__(self, idx, value):
if idx < 0:
idx += len(self)
if idx < len(self):
_tsubasa.Slice_uint8_set(self.handle, idx, value)
return
raise IndexError('slice index out of range')
def __iadd__(self, value):
if not isinstance(value, _collections_abc.Iterable):
raise TypeError('Slice_uint8.__iadd__ takes a sequence as argument')
for elt in value:
self.append(elt)
return self
def __iter__(self):
self.index = 0
return self
def __next__(self):
if self.index < len(self):
rv = _tsubasa.Slice_uint8_elem(self.handle, self.index)
self.index = self.index + 1
return rv
raise StopIteration
def append(self, value):
_tsubasa.Slice_uint8_append(self.handle, value)
def copy(self, src):
""" copy emulates the go copy function, copying elements into this list from source list, up to min of size of each list """
mx = min(len(self), len(src))
for i in range(mx):
self[i] = src[i]
# ---- External Types Outside of Targeted Packages ---
| 32.59366 | 126 | 0.676415 | 6,844 | 45,240 | 4.267972 | 0.029369 | 0.087984 | 0.027936 | 0.037795 | 0.930503 | 0.924478 | 0.906265 | 0.900719 | 0.900719 | 0.900719 | 0 | 0.015391 | 0.191424 | 45,240 | 1,387 | 127 | 32.617159 | 0.783133 | 0.114478 | 0 | 0.817959 | 1 | 0 | 0.095411 | 0.006462 | 0 | 0 | 0 | 0 | 0 | 1 | 0.159184 | false | 0 | 0.005714 | 0.026122 | 0.309388 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 |
eae1bb293c1312b37011e03240675835981b85d2 | 121 | py | Python | Hello world.py | KangFrank/Python_Start | 5d09ef97dd3960d0cb1b99dca017fbc288dbc98c | [
"FSFAP"
] | 2 | 2017-07-26T07:57:55.000Z | 2018-09-30T14:39:21.000Z | Hello world.py | KangFrank/Python_Start | 5d09ef97dd3960d0cb1b99dca017fbc288dbc98c | [
"FSFAP"
] | null | null | null | Hello world.py | KangFrank/Python_Start | 5d09ef97dd3960d0cb1b99dca017fbc288dbc98c | [
"FSFAP"
] | null | null | null | #-*-coding:UTF-8-*-
FileName:Hello world.py
#python 2.6/2.7
print "Hello world.\n"
#python 3.6
print("Hello world.\n")
| 13.444444 | 23 | 0.661157 | 23 | 121 | 3.478261 | 0.608696 | 0.375 | 0.375 | 0.4 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.065421 | 0.115702 | 121 | 8 | 24 | 15.125 | 0.682243 | 0.347107 | 0 | 0 | 0 | 0 | 0.368421 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | null | 0 | 0 | null | null | 0.666667 | 1 | 0 | 0 | null | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 7 |
d80f7ebbe37a732fe098e30ffb39ce28cc434721 | 3,201 | py | Python | tests/test_lib.py | DK26/pyrust-keyring | f3a5aecf24c2ba7fbcf28e0d54eb0ef249c8fe1d | [
"Apache-2.0",
"MIT"
] | 1 | 2020-06-09T19:11:13.000Z | 2020-06-09T19:11:13.000Z | tests/test_lib.py | DK26/pyrust-keyring | f3a5aecf24c2ba7fbcf28e0d54eb0ef249c8fe1d | [
"Apache-2.0",
"MIT"
] | null | null | null | tests/test_lib.py | DK26/pyrust-keyring | f3a5aecf24c2ba7fbcf28e0d54eb0ef249c8fe1d | [
"Apache-2.0",
"MIT"
] | null | null | null | import unittest
from uuid import uuid4
import rskeyring
def generate_password():
return str(uuid4())
class RSKeyringTests(unittest.TestCase):
service_name = "rskeyring_unit_test_service_name"
username = "rskeyring_unit_test_username"
def test_set_password(self):
rskeyring.set_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username,
password=generate_password()
)
def test_update_password(self):
first_password = generate_password()
rskeyring.set_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username,
password=first_password
)
stored_password = rskeyring.get_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username
)
self.assertEqual(first_password, stored_password)
second_password = generate_password()
rskeyring.set_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username,
password=second_password
)
stored_password = rskeyring.get_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username
)
self.assertEqual(second_password, stored_password)
def test_get_password(self):
generated_password = generate_password()
rskeyring.set_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username,
password=generated_password
)
stored_password = rskeyring.get_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username
)
self.assertEqual(stored_password, generated_password)
def test_delete_password(self):
generated_password = generate_password()
rskeyring.set_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username,
password=generated_password
)
stored_password = rskeyring.get_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username
)
self.assertEqual(stored_password, generated_password)
rskeyring.delete_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username
)
with self.assertRaises(OSError):
rskeyring.get_password(
service=RSKeyringTests.service_name,
username=RSKeyringTests.username
)
def test_exceptions(self):
random_user_name = 'rskeyring_unittest_' + generate_password()
with self.assertRaises(OSError):
rskeyring.get_password(
service=RSKeyringTests.service_name,
username=random_user_name
)
with self.assertRaises(OSError):
rskeyring.delete_password(
service=RSKeyringTests.service_name,
username=random_user_name
)
if __name__ == '__main__':
unittest.main()
| 26.454545 | 70 | 0.646673 | 273 | 3,201 | 7.267399 | 0.135531 | 0.083165 | 0.134073 | 0.235887 | 0.777722 | 0.764113 | 0.764113 | 0.764113 | 0.71371 | 0.675403 | 0 | 0.00088 | 0.289597 | 3,201 | 120 | 71 | 26.675 | 0.871592 | 0 | 0 | 0.571429 | 1 | 0 | 0.027179 | 0.018744 | 0 | 0 | 0 | 0 | 0.083333 | 1 | 0.071429 | false | 0.380952 | 0.035714 | 0.011905 | 0.154762 | 0 | 0 | 0 | 0 | null | 0 | 0 | 1 | 0 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 8 |
d823e7061d051ef86b50d04b6b3e4bc08cba9ba2 | 73 | py | Python | test/test.py | amraf1002/my-first-project | 49d4af2e225b358ba7d2a4f97f78b8a208597ee7 | [
"MIT"
] | null | null | null | test/test.py | amraf1002/my-first-project | 49d4af2e225b358ba7d2a4f97f78b8a208597ee7 | [
"MIT"
] | null | null | null | test/test.py | amraf1002/my-first-project | 49d4af2e225b358ba7d2a4f97f78b8a208597ee7 | [
"MIT"
] | null | null | null | from awesome_utility import hello
print(hello())
print(hello('Frank'))
| 12.166667 | 33 | 0.753425 | 10 | 73 | 5.4 | 0.7 | 0.37037 | 0.555556 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.109589 | 73 | 5 | 34 | 14.6 | 0.830769 | 0 | 0 | 0 | 0 | 0 | 0.068493 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 0.333333 | 0 | 0.333333 | 0.666667 | 1 | 0 | 0 | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 0 | 1 | 0 | 7 |
dc91c974ed98e20cdb5647d9665f79707049b4f9 | 141 | py | Python | pandas_ml/core/__init__.py | matsavage/pandas-ml | 794cddc8dc5d0a49fbc9734d826d9465078f376e | [
"BSD-3-Clause"
] | 305 | 2016-02-21T06:35:25.000Z | 2022-03-30T11:53:31.000Z | pandas_ml/core/__init__.py | matsavage/pandas-ml | 794cddc8dc5d0a49fbc9734d826d9465078f376e | [
"BSD-3-Clause"
] | 69 | 2016-02-16T08:10:46.000Z | 2022-03-04T14:36:12.000Z | pandas_ml/core/__init__.py | matsavage/pandas-ml | 794cddc8dc5d0a49fbc9734d826d9465078f376e | [
"BSD-3-Clause"
] | 73 | 2016-02-16T08:27:28.000Z | 2022-03-10T06:57:51.000Z | #!/usr/bin/env python
from pandas_ml.core.frame import ModelFrame # noqa
from pandas_ml.core.series import ModelSeries # noqa
| 28.2 | 57 | 0.716312 | 20 | 141 | 4.95 | 0.7 | 0.20202 | 0.242424 | 0.323232 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.205674 | 141 | 4 | 58 | 35.25 | 0.883929 | 0.212766 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 7 |
762db5e829492f610dce8dbc5dd54677af4b0814 | 56,407 | py | Python | google/cloud/channel_v1/services/cloud_channel_service/pagers.py | renovate-bot/python-channel | 0cebe664e7a5889269f63237f7c365656299e5d7 | [
"Apache-2.0"
] | 1 | 2021-11-13T12:15:04.000Z | 2021-11-13T12:15:04.000Z | google/cloud/channel_v1/services/cloud_channel_service/pagers.py | renovate-bot/python-channel | 0cebe664e7a5889269f63237f7c365656299e5d7 | [
"Apache-2.0"
] | 45 | 2021-01-14T23:02:15.000Z | 2022-03-08T01:01:52.000Z | google/cloud/channel_v1/services/cloud_channel_service/pagers.py | renovate-bot/python-channel | 0cebe664e7a5889269f63237f7c365656299e5d7 | [
"Apache-2.0"
] | 2 | 2021-01-14T22:39:56.000Z | 2022-01-29T08:09:14.000Z | # -*- coding: utf-8 -*-
# Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
from typing import (
Any,
AsyncIterator,
Awaitable,
Callable,
Sequence,
Tuple,
Optional,
Iterator,
)
from google.cloud.channel_v1.types import channel_partner_links
from google.cloud.channel_v1.types import customers
from google.cloud.channel_v1.types import entitlements
from google.cloud.channel_v1.types import offers
from google.cloud.channel_v1.types import products
from google.cloud.channel_v1.types import service
class ListCustomersPager:
"""A pager for iterating through ``list_customers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListCustomersResponse` object, and
provides an ``__iter__`` method to iterate through its
``customers`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListCustomers`` requests and continue to iterate
through the ``customers`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListCustomersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListCustomersResponse],
request: service.ListCustomersRequest,
response: service.ListCustomersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListCustomersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListCustomersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListCustomersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListCustomersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[customers.Customer]:
for page in self.pages:
yield from page.customers
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListCustomersAsyncPager:
"""A pager for iterating through ``list_customers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListCustomersResponse` object, and
provides an ``__aiter__`` method to iterate through its
``customers`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListCustomers`` requests and continue to iterate
through the ``customers`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListCustomersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListCustomersResponse]],
request: service.ListCustomersRequest,
response: service.ListCustomersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListCustomersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListCustomersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListCustomersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListCustomersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[customers.Customer]:
async def async_generator():
async for page in self.pages:
for response in page.customers:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListEntitlementsPager:
"""A pager for iterating through ``list_entitlements`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListEntitlementsResponse` object, and
provides an ``__iter__`` method to iterate through its
``entitlements`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListEntitlements`` requests and continue to iterate
through the ``entitlements`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListEntitlementsResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListEntitlementsResponse],
request: service.ListEntitlementsRequest,
response: service.ListEntitlementsResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListEntitlementsRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListEntitlementsResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListEntitlementsRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListEntitlementsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[entitlements.Entitlement]:
for page in self.pages:
yield from page.entitlements
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListEntitlementsAsyncPager:
"""A pager for iterating through ``list_entitlements`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListEntitlementsResponse` object, and
provides an ``__aiter__`` method to iterate through its
``entitlements`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListEntitlements`` requests and continue to iterate
through the ``entitlements`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListEntitlementsResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListEntitlementsResponse]],
request: service.ListEntitlementsRequest,
response: service.ListEntitlementsResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListEntitlementsRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListEntitlementsResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListEntitlementsRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListEntitlementsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[entitlements.Entitlement]:
async def async_generator():
async for page in self.pages:
for response in page.entitlements:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListTransferableSkusPager:
"""A pager for iterating through ``list_transferable_skus`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListTransferableSkusResponse` object, and
provides an ``__iter__`` method to iterate through its
``transferable_skus`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListTransferableSkus`` requests and continue to iterate
through the ``transferable_skus`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListTransferableSkusResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListTransferableSkusResponse],
request: service.ListTransferableSkusRequest,
response: service.ListTransferableSkusResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListTransferableSkusRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListTransferableSkusResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListTransferableSkusRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListTransferableSkusResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[entitlements.TransferableSku]:
for page in self.pages:
yield from page.transferable_skus
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListTransferableSkusAsyncPager:
"""A pager for iterating through ``list_transferable_skus`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListTransferableSkusResponse` object, and
provides an ``__aiter__`` method to iterate through its
``transferable_skus`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListTransferableSkus`` requests and continue to iterate
through the ``transferable_skus`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListTransferableSkusResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListTransferableSkusResponse]],
request: service.ListTransferableSkusRequest,
response: service.ListTransferableSkusResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListTransferableSkusRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListTransferableSkusResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListTransferableSkusRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListTransferableSkusResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[entitlements.TransferableSku]:
async def async_generator():
async for page in self.pages:
for response in page.transferable_skus:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListTransferableOffersPager:
"""A pager for iterating through ``list_transferable_offers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListTransferableOffersResponse` object, and
provides an ``__iter__`` method to iterate through its
``transferable_offers`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListTransferableOffers`` requests and continue to iterate
through the ``transferable_offers`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListTransferableOffersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListTransferableOffersResponse],
request: service.ListTransferableOffersRequest,
response: service.ListTransferableOffersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListTransferableOffersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListTransferableOffersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListTransferableOffersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListTransferableOffersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[service.TransferableOffer]:
for page in self.pages:
yield from page.transferable_offers
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListTransferableOffersAsyncPager:
"""A pager for iterating through ``list_transferable_offers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListTransferableOffersResponse` object, and
provides an ``__aiter__`` method to iterate through its
``transferable_offers`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListTransferableOffers`` requests and continue to iterate
through the ``transferable_offers`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListTransferableOffersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListTransferableOffersResponse]],
request: service.ListTransferableOffersRequest,
response: service.ListTransferableOffersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListTransferableOffersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListTransferableOffersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListTransferableOffersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListTransferableOffersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[service.TransferableOffer]:
async def async_generator():
async for page in self.pages:
for response in page.transferable_offers:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListChannelPartnerLinksPager:
"""A pager for iterating through ``list_channel_partner_links`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListChannelPartnerLinksResponse` object, and
provides an ``__iter__`` method to iterate through its
``channel_partner_links`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListChannelPartnerLinks`` requests and continue to iterate
through the ``channel_partner_links`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListChannelPartnerLinksResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListChannelPartnerLinksResponse],
request: service.ListChannelPartnerLinksRequest,
response: service.ListChannelPartnerLinksResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListChannelPartnerLinksRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListChannelPartnerLinksResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListChannelPartnerLinksRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListChannelPartnerLinksResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[channel_partner_links.ChannelPartnerLink]:
for page in self.pages:
yield from page.channel_partner_links
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListChannelPartnerLinksAsyncPager:
"""A pager for iterating through ``list_channel_partner_links`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListChannelPartnerLinksResponse` object, and
provides an ``__aiter__`` method to iterate through its
``channel_partner_links`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListChannelPartnerLinks`` requests and continue to iterate
through the ``channel_partner_links`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListChannelPartnerLinksResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListChannelPartnerLinksResponse]],
request: service.ListChannelPartnerLinksRequest,
response: service.ListChannelPartnerLinksResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListChannelPartnerLinksRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListChannelPartnerLinksResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListChannelPartnerLinksRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListChannelPartnerLinksResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[channel_partner_links.ChannelPartnerLink]:
async def async_generator():
async for page in self.pages:
for response in page.channel_partner_links:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListProductsPager:
"""A pager for iterating through ``list_products`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListProductsResponse` object, and
provides an ``__iter__`` method to iterate through its
``products`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListProducts`` requests and continue to iterate
through the ``products`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListProductsResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListProductsResponse],
request: service.ListProductsRequest,
response: service.ListProductsResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListProductsRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListProductsResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListProductsRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListProductsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[products.Product]:
for page in self.pages:
yield from page.products
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListProductsAsyncPager:
"""A pager for iterating through ``list_products`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListProductsResponse` object, and
provides an ``__aiter__`` method to iterate through its
``products`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListProducts`` requests and continue to iterate
through the ``products`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListProductsResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListProductsResponse]],
request: service.ListProductsRequest,
response: service.ListProductsResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListProductsRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListProductsResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListProductsRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListProductsResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[products.Product]:
async def async_generator():
async for page in self.pages:
for response in page.products:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListSkusPager:
"""A pager for iterating through ``list_skus`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListSkusResponse` object, and
provides an ``__iter__`` method to iterate through its
``skus`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListSkus`` requests and continue to iterate
through the ``skus`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListSkusResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListSkusResponse],
request: service.ListSkusRequest,
response: service.ListSkusResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListSkusRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListSkusResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListSkusRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListSkusResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[products.Sku]:
for page in self.pages:
yield from page.skus
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListSkusAsyncPager:
"""A pager for iterating through ``list_skus`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListSkusResponse` object, and
provides an ``__aiter__`` method to iterate through its
``skus`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListSkus`` requests and continue to iterate
through the ``skus`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListSkusResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListSkusResponse]],
request: service.ListSkusRequest,
response: service.ListSkusResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListSkusRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListSkusResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListSkusRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListSkusResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[products.Sku]:
async def async_generator():
async for page in self.pages:
for response in page.skus:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListOffersPager:
"""A pager for iterating through ``list_offers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListOffersResponse` object, and
provides an ``__iter__`` method to iterate through its
``offers`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListOffers`` requests and continue to iterate
through the ``offers`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListOffersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListOffersResponse],
request: service.ListOffersRequest,
response: service.ListOffersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListOffersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListOffersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListOffersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListOffersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[offers.Offer]:
for page in self.pages:
yield from page.offers
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListOffersAsyncPager:
"""A pager for iterating through ``list_offers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListOffersResponse` object, and
provides an ``__aiter__`` method to iterate through its
``offers`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListOffers`` requests and continue to iterate
through the ``offers`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListOffersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListOffersResponse]],
request: service.ListOffersRequest,
response: service.ListOffersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListOffersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListOffersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListOffersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListOffersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[offers.Offer]:
async def async_generator():
async for page in self.pages:
for response in page.offers:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListPurchasableSkusPager:
"""A pager for iterating through ``list_purchasable_skus`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListPurchasableSkusResponse` object, and
provides an ``__iter__`` method to iterate through its
``purchasable_skus`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListPurchasableSkus`` requests and continue to iterate
through the ``purchasable_skus`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListPurchasableSkusResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListPurchasableSkusResponse],
request: service.ListPurchasableSkusRequest,
response: service.ListPurchasableSkusResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListPurchasableSkusRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListPurchasableSkusResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListPurchasableSkusRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListPurchasableSkusResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[service.PurchasableSku]:
for page in self.pages:
yield from page.purchasable_skus
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListPurchasableSkusAsyncPager:
"""A pager for iterating through ``list_purchasable_skus`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListPurchasableSkusResponse` object, and
provides an ``__aiter__`` method to iterate through its
``purchasable_skus`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListPurchasableSkus`` requests and continue to iterate
through the ``purchasable_skus`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListPurchasableSkusResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListPurchasableSkusResponse]],
request: service.ListPurchasableSkusRequest,
response: service.ListPurchasableSkusResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListPurchasableSkusRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListPurchasableSkusResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListPurchasableSkusRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListPurchasableSkusResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[service.PurchasableSku]:
async def async_generator():
async for page in self.pages:
for response in page.purchasable_skus:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListPurchasableOffersPager:
"""A pager for iterating through ``list_purchasable_offers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListPurchasableOffersResponse` object, and
provides an ``__iter__`` method to iterate through its
``purchasable_offers`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListPurchasableOffers`` requests and continue to iterate
through the ``purchasable_offers`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListPurchasableOffersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListPurchasableOffersResponse],
request: service.ListPurchasableOffersRequest,
response: service.ListPurchasableOffersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListPurchasableOffersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListPurchasableOffersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListPurchasableOffersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListPurchasableOffersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[service.PurchasableOffer]:
for page in self.pages:
yield from page.purchasable_offers
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListPurchasableOffersAsyncPager:
"""A pager for iterating through ``list_purchasable_offers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListPurchasableOffersResponse` object, and
provides an ``__aiter__`` method to iterate through its
``purchasable_offers`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListPurchasableOffers`` requests and continue to iterate
through the ``purchasable_offers`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListPurchasableOffersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListPurchasableOffersResponse]],
request: service.ListPurchasableOffersRequest,
response: service.ListPurchasableOffersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListPurchasableOffersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListPurchasableOffersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListPurchasableOffersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListPurchasableOffersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[service.PurchasableOffer]:
async def async_generator():
async for page in self.pages:
for response in page.purchasable_offers:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListSubscribersPager:
"""A pager for iterating through ``list_subscribers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListSubscribersResponse` object, and
provides an ``__iter__`` method to iterate through its
``service_accounts`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListSubscribers`` requests and continue to iterate
through the ``service_accounts`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListSubscribersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., service.ListSubscribersResponse],
request: service.ListSubscribersRequest,
response: service.ListSubscribersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListSubscribersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListSubscribersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListSubscribersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterator[service.ListSubscribersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterator[str]:
for page in self.pages:
yield from page.service_accounts
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
class ListSubscribersAsyncPager:
"""A pager for iterating through ``list_subscribers`` requests.
This class thinly wraps an initial
:class:`google.cloud.channel_v1.types.ListSubscribersResponse` object, and
provides an ``__aiter__`` method to iterate through its
``service_accounts`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListSubscribers`` requests and continue to iterate
through the ``service_accounts`` field on the
corresponding responses.
All the usual :class:`google.cloud.channel_v1.types.ListSubscribersResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(
self,
method: Callable[..., Awaitable[service.ListSubscribersResponse]],
request: service.ListSubscribersRequest,
response: service.ListSubscribersResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()
):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.channel_v1.types.ListSubscribersRequest):
The initial request object.
response (google.cloud.channel_v1.types.ListSubscribersResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = service.ListSubscribersRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterator[service.ListSubscribersResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterator[str]:
async def async_generator():
async for page in self.pages:
for response in page.service_accounts:
yield response
return async_generator()
def __repr__(self) -> str:
return "{0}<{1!r}>".format(self.__class__.__name__, self._response)
| 39.144344 | 88 | 0.670715 | 6,040 | 56,407 | 6.041556 | 0.039073 | 0.057877 | 0.046368 | 0.05152 | 0.957716 | 0.957716 | 0.957716 | 0.951961 | 0.946014 | 0.941355 | 0 | 0.003457 | 0.246086 | 56,407 | 1,440 | 89 | 39.171528 | 0.85463 | 0.455599 | 0 | 0.804044 | 0 | 0 | 0.007967 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.153966 | false | 0 | 0.010886 | 0.068429 | 0.284603 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
5201e84107f68ae07658aab43200f274c793872c | 3,834 | py | Python | forms.py | StarkerRegen/Pixelboooom | 5b8239d66e9a9f4c890472f90d9fa9ffc593aa00 | [
"MIT"
] | null | null | null | forms.py | StarkerRegen/Pixelboooom | 5b8239d66e9a9f4c890472f90d9fa9ffc593aa00 | [
"MIT"
] | 3 | 2021-06-08T22:18:26.000Z | 2022-03-12T00:46:44.000Z | forms.py | StarkerRegen/Pixelboooom | 5b8239d66e9a9f4c890472f90d9fa9ffc593aa00 | [
"MIT"
] | null | null | null | from flask_wtf import FlaskForm
from wtforms import StringField, BooleanField, PasswordField, SubmitField, IntegerField, TextAreaField
from wtforms.validators import DataRequired, Length, Email, Regexp, EqualTo, ValidationError
from models import User
class SigninForm(FlaskForm):
# 域初始化时,第一个参数是设置label属性的
email = StringField('EmailAddress', validators=[DataRequired(), Email(message='Email address error.')])
password = PasswordField('Password', validators=[DataRequired(), Length(8,16),
Regexp('^[a-zA-Z0-9]*$', message="The password should contain only a-z, A-z and 0-9." )])
remember_me = BooleanField('RememberMe', default=False)
submit = SubmitField('Sign in')
class SignupForm(FlaskForm):
username = StringField('UserName', validators=[DataRequired(), Length(1,20),
Regexp('^[a-zA-Z0-9]*$', message="The username should contain only a-z, A-z and 0-9." )])
email = StringField('EmailAddress', validators=[DataRequired(), Email(message='Email address error.')])
password = PasswordField('Password', validators=[DataRequired(), Length(8,16),
Regexp('^[a-zA-Z0-9]*$', message="The password should contain only a-z, A-z and 0-9." )])
password_c = PasswordField('Password_c', validators=[DataRequired(), EqualTo('password',message='Password error')])
submit = SubmitField('Create account')
def validate_username(self, username):
user = User.query.filter_by(username=username.data).first()
if user:
raise ValidationError('User already exist.')
def validate_email(self, email):
user = User.query.filter_by(email=email.data).first()
if user:
raise ValidationError('Email has been registered, please re-enter')
class ResetForm(FlaskForm):
username = StringField('UserName', validators=[DataRequired(), Length(1,20),
Regexp('^[a-zA-Z0-9]*$', message="The username should contain only a-z, A-z and 0-9." )])
email = StringField('EmailAddress', validators=[DataRequired(), Email(message='Email address error.')])
password = PasswordField('Password', validators=[DataRequired(), Length(8,16),
Regexp('^[a-zA-Z0-9]*$', message="The password should contain only a-z, A-z and 0-9." )])
password_c = PasswordField('Password_c', validators=[DataRequired(), EqualTo('password',message='Password error')])
submit = SubmitField('Reset Passwd')
def validate_username(self, username):
user = User.query.filter_by(username=username.data).first()
if not(user):
raise ValidationError("User not registered.")
def validate_email(self, email):
user = User.query.filter_by(email=email.data).first()
if not(user):
raise ValidationError('User not registered.')
class PostForm(FlaskForm):
title = StringField('title', validators=[DataRequired(), Length(1,20,message='Length out of limit')])
category = StringField('category', validators=[DataRequired(), Length(1,50,message='Length out of limit')])
img = StringField('id', validators=[DataRequired(), Length(1,40,message='Length out of limit')])
style = StringField('style', validators=[DataRequired(), Length(1,10,message='Length out of limit')])
article = TextAreaField('style', validators=[DataRequired(), Length(1,120,message='Length out of limit')])
submit = SubmitField('Publish')
class EditForm(FlaskForm):
title = StringField('title', validators=[DataRequired(), Length(1,20,message='Length out of limit')])
category = StringField('category', validators=[DataRequired(), Length(1,50,message='Length out of limit')])
article = TextAreaField('style', validators=[DataRequired(), Length(1,120,message='Length out of limit')])
submit = SubmitField('Update')
| 58.984615 | 119 | 0.68362 | 447 | 3,834 | 5.832215 | 0.217002 | 0.151899 | 0.139624 | 0.111239 | 0.795167 | 0.773303 | 0.754891 | 0.754891 | 0.754891 | 0.754891 | 0 | 0.01908 | 0.166145 | 3,834 | 65 | 120 | 58.984615 | 0.796372 | 0.005738 | 0 | 0.611111 | 0 | 0 | 0.228811 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.074074 | false | 0.185185 | 0.074074 | 0 | 0.685185 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 7 |
5204fbad42cee1a49426d633df4f39a7e5239455 | 139 | py | Python | app/monitor/__init__.py | bcarroll/inmoov_mini | 029152cb186d45a040ddc71fd562191bff82b30c | [
"MIT"
] | null | null | null | app/monitor/__init__.py | bcarroll/inmoov_mini | 029152cb186d45a040ddc71fd562191bff82b30c | [
"MIT"
] | null | null | null | app/monitor/__init__.py | bcarroll/inmoov_mini | 029152cb186d45a040ddc71fd562191bff82b30c | [
"MIT"
] | null | null | null | from flask import Blueprint
bp = Blueprint('monitor', __name__)
from app.monitor import routes_memory
from app.monitor import routes_cpu
| 19.857143 | 37 | 0.81295 | 20 | 139 | 5.35 | 0.55 | 0.130841 | 0.261682 | 0.373832 | 0.485981 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.129496 | 139 | 6 | 38 | 23.166667 | 0.884298 | 0 | 0 | 0 | 0 | 0 | 0.05036 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.75 | 0 | 0.75 | 0.5 | 1 | 0 | 0 | null | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 1 | 0 | 7 |
52273f979806993100f37ec772519c92c772855f | 1,799 | py | Python | src/models.py | ryzenboi98/btc-data-api | e9d7447595a35834bf63e738e797fb8b26c9a134 | [
"MIT"
] | null | null | null | src/models.py | ryzenboi98/btc-data-api | e9d7447595a35834bf63e738e797fb8b26c9a134 | [
"MIT"
] | null | null | null | src/models.py | ryzenboi98/btc-data-api | e9d7447595a35834bf63e738e797fb8b26c9a134 | [
"MIT"
] | null | null | null | from sqlalchemy import Integer, String, Float, DateTime
from sqlalchemy.sql.expression import null
from sqlalchemy.sql.schema import Column
from src.database import Base
class Stats(Base):
__tablename__ = 'stats'
id = Column(Integer, primary_key=True)
lower = Column(Float, nullable=False)
higher = Column(Float, nullable=False)
open = Column(Float, nullable=False)
close = Column(Float, nullable=False)
volume = Column(Float, nullable=False)
change = Column(Float, nullable=True)
timestamp = Column(DateTime, nullable=False)
class WeeklyStats(Base):
__tablename__ = 'weekly_stats'
id = Column(Integer, primary_key=True)
lower = Column(Float, nullable=False)
higher = Column(Float, nullable=False)
open = Column(Float, nullable=False)
close = Column(Float, nullable=False)
avg_volume = Column(Float, nullable=False)
change = Column(Float, nullable=True)
timestamp = Column(DateTime, nullable=False)
class MonthlyStats(Base):
__tablename__ = 'monthly_stats'
id = Column(Integer, primary_key=True)
lower = Column(Float, nullable=False)
higher = Column(Float, nullable=False)
open = Column(Float, nullable=False)
close = Column(Float, nullable=False)
avg_volume = Column(Float, nullable=False)
change = Column(Float, nullable=True)
timestamp = Column(DateTime, nullable=False)
class AnualStats(Base):
__tablename__ = 'anual_stats'
id = Column(Integer, primary_key=True)
lower = Column(Float, nullable=False)
higher = Column(Float, nullable=False)
open = Column(Float, nullable=False)
close = Column(Float, nullable=False)
avg_volume = Column(Float, nullable=False)
change = Column(Float, nullable=True)
timestamp = Column(DateTime, nullable=False)
| 32.709091 | 55 | 0.714842 | 214 | 1,799 | 5.88785 | 0.17757 | 0.209524 | 0.361905 | 0.380952 | 0.796825 | 0.796825 | 0.796825 | 0.796825 | 0.796825 | 0.796825 | 0 | 0 | 0.178432 | 1,799 | 54 | 56 | 33.314815 | 0.852503 | 0 | 0 | 0.704545 | 0 | 0 | 0.022803 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.090909 | 0 | 1 | 0 | 0 | 0 | 0 | null | 1 | 1 | 1 | 0 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 9 |
523adcb3b588cbbb6467966415118284722b39b4 | 3,641 | py | Python | snipeit/Fields.py | adfedotov/snipeit-api | 0f74a4a1957f6e1bd247af714309d429d38c7331 | [
"MIT"
] | null | null | null | snipeit/Fields.py | adfedotov/snipeit-api | 0f74a4a1957f6e1bd247af714309d429d38c7331 | [
"MIT"
] | null | null | null | snipeit/Fields.py | adfedotov/snipeit-api | 0f74a4a1957f6e1bd247af714309d429d38c7331 | [
"MIT"
] | null | null | null | import requests as r
class Fields():
def __init__(self, server, headers):
self.server = server
self.headers = headers
def get(self):
"""Get custom fields
Returns:
dict: Json response
"""
endpoint = self.server + '/api/v1/fields'
response = r.request('GET', endpoint, headers=self.headers)
return response.json()
def get_field_by_id(self, id: int):
"""Get field by id
Args:
id (int): field id
Returns:
dict: Json response
"""
endpoint = f'{self.server}/api/v1/fields/{id}'
response = r.request('GET', endpoint, headers=self.headers)
return response.json()
def create(self, payload: dict):
"""Create field
Args:
payload (dict): Payload has to contain 'name', 'element'. Accepted:
# name (str)
# element (str) [text, textarea, checkbox, radio, listbox]
# field_values (str)
# show_in_email (bool)
# format (str)
# field_encrypted (bool)
# help_text (str)
Returns:
dict: Json response
"""
endpoint = f'{self.server}/api/v1/fields'
response = r.request('POST', endpoint, headers=self.headers, json=payload)
return response.json()
def update(self, id: int, payload: dict):
"""Update field
Args:
id (int): field ID
Payload has to contain 'name', 'element'. Accepted:
# name (str)
# element (str) [text, textarea, checkbox, radio, listbox]
# field_values (str)
# show_in_email (bool)
# format (str)
# field_encrypted (bool)
# help_text (str)
Returns:
dict: Json response
"""
endpoint = f'{self.server}/api/v1/fields/{id}'
response = r.request('PUT', endpoint, headers=self.headers, json=payload)
return response.json()
def associate_with(self, id: int, fieldset_id: int):
"""Associate custom field with a custom fieldset
Args:
id (int): field id
fieldset_id (int): fieldset id
Returns:
dict: Json response
"""
payload = {
'fieldset_id': fieldset_id
}
endpoint = f'{self.server}/api/v1/fields/{id}/associate'
response = r.request('POST', endpoint, headers=self.headers, json=payload)
return response.json()
def disassociate_with(self, id: int, fieldset_id: int):
"""Disassociate custom field with a custom fieldset
Args:
id (int): field id
fieldset_id (int): fieldset id
Returns:
dict: Json response
"""
payload = {
'fieldset_id': fieldset_id
}
endpoint = f'{self.server}/api/v1/fields/{id}/disassociate'
response = r.request('POST', endpoint, headers=self.headers, json=payload)
return response.json()
def delete(self, id: int):
"""Delete field
Args:
id (int): field id
Returns:
dict: Json response
"""
endpoint = f'{self.server}/api/v1/fields/{id}'
response = r.request('DELETE', endpoint, headers=self.headers)
return response.json()
| 27.171642 | 86 | 0.503433 | 368 | 3,641 | 4.907609 | 0.160326 | 0.03876 | 0.05814 | 0.089147 | 0.825028 | 0.807863 | 0.793466 | 0.74031 | 0.728682 | 0.728682 | 0 | 0.003142 | 0.38808 | 3,641 | 134 | 87 | 27.171642 | 0.807451 | 0.381214 | 0 | 0.487179 | 0 | 0 | 0.150745 | 0.115958 | 0 | 0 | 0 | 0 | 0 | 1 | 0.205128 | false | 0 | 0.025641 | 0 | 0.435897 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
524b02c3b4cfe186f11492b2be2e296ebfec1a32 | 22,094 | py | Python | tests/integration/views/surveys/help/test_surveys_help_info_about_this_survey.py | ONSdigital/ras-frontstage | e8ff1931b49cb3ab47b421aed6780e9e944dceea | [
"MIT"
] | 8 | 2017-06-30T12:32:02.000Z | 2022-02-25T09:07:28.000Z | tests/integration/views/surveys/help/test_surveys_help_info_about_this_survey.py | ONSdigital/ras-frontstage | e8ff1931b49cb3ab47b421aed6780e9e944dceea | [
"MIT"
] | 256 | 2017-05-16T09:38:09.000Z | 2022-03-28T13:38:42.000Z | tests/integration/views/surveys/help/test_surveys_help_info_about_this_survey.py | ONSdigital/ras-frontstage | e8ff1931b49cb3ab47b421aed6780e9e944dceea | [
"MIT"
] | 4 | 2017-09-29T08:58:36.000Z | 2021-04-11T07:44:27.000Z | import unittest
from unittest.mock import patch
import requests_mock
from frontstage import app
from tests.integration.mocked_services import (
business_party,
encoded_jwt_token,
respondent_party,
survey,
survey_eq,
survey_list_todo,
survey_rsi,
url_banner_api,
)
class TestSurveyHelpInfoAboutThisSurvey(unittest.TestCase):
def setUp(self):
self.app = app.test_client()
self.app.set_cookie("localhost", "authorization", "session_key")
self.headers = {
"Authorization": "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJ1c2VyX2lkIjoicmluZ3JhbUBub3d3aGVyZS5jb20iLCJ1c2Vy"
"X3Njb3BlcyI6WyJjaS5yZWFkIiwiY2kud3JpdGUiXX0.se0BJtNksVtk14aqjp7SvnXzRbEKoqXb8Q5U9VVdy54"
# NOQA
}
self.patcher = patch("redis.StrictRedis.get", return_value=encoded_jwt_token)
self.patcher.start()
def tearDown(self):
self.patcher.stop()
def set_flask_session(self):
with self.app.session_transaction() as mock_session:
mock_session["help_survey_ref"] = "074"
mock_session["help_ru_ref"] = "49900000001F"
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_qbs(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey_eq
get_business.return_value = business_party
response = self.app.get("/surveys/surveys-help?survey_ref=139&ru_ref=49900000001F", follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Help".encode(), response.data)
self.assertIn("Choose an option".encode(), response.data)
self.assertIn("Information about the Quarterly Business Survey".encode(), response.data)
self.assertIn("Continue".encode(), response.data)
self.assertIn("Cancel".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_bricks(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
response = self.app.get("/surveys/surveys-help?survey_ref=074&ru_ref=49900000001F", follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Help".encode(), response.data)
self.assertIn("Choose an option".encode(), response.data)
self.assertIn("Information about the Monthly Survey of Building Materials Bricks".encode(), response.data)
self.assertIn("Continue".encode(), response.data)
self.assertIn("Cancel".encode(), response.data)
self.assertIn("Help".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_bricks_with_no_option_select(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
form = {}
self.set_flask_session()
response = self.app.post("/surveys/help", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("There is 1 error on this page".encode(), response.data)
self.assertIn("You need to choose an option".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_bricks_with_option_select(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
form = {"option": "info-about-this-survey"}
self.set_flask_session()
response = self.app.post("/surveys/help", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Information about the Monthly Survey of Building Materials Bricks".encode(), response.data)
self.assertIn("Choose an option".encode(), response.data)
self.assertIn("Can I be exempt from completing the survey questionnaire?".encode(), response.data)
self.assertIn("How was my business selected?".encode(), response.data)
self.assertIn("How long will it take to complete?".encode(), response.data)
self.assertIn("How long will my business be selected for?".encode(), response.data)
self.assertIn("Something else".encode(), response.data)
self.assertIn("Continue".encode(), response.data)
self.assertIn("Cancel".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_bricks_with_sub_option_exemption_completing_survey(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
form = {"option": "exemption-completing-survey"}
self.set_flask_session()
response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Can I be exempt from completing the survey questionnaire?".encode(), response.data)
self.assertIn(
"While this survey is voluntary, we have selected your company in the same way we ".encode(), response.data
)
self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data)
self.assertIn("Did this answer your question?".encode(), response.data)
self.assertIn("Yes".encode(), response.data)
self.assertIn("No".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_rsi_with_sub_option_exemption_completing_survey(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey_rsi
get_business.return_value = business_party
form = {"option": "exemption-completing-survey"}
self.set_flask_session()
response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Can I be exempt from completing the survey questionnaire?".encode(), response.data)
self.assertIn(
"No. Once selected, the law obliges a business to complete the survey under the provisions "
"of the Statistics of Trade Act 1947".encode(),
response.data,
)
self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data)
self.assertIn("Did this answer your question?".encode(), response.data)
self.assertIn("Yes".encode(), response.data)
self.assertIn("No".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_bricks_with_sub_option_why_selected(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
form = {"option": "why-selected"}
self.set_flask_session()
response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("How was my business selected?".encode(), response.data)
self.assertIn(
"We select businesses from the Inter-Departmental Business Register (IDBR). ".encode(), response.data
)
self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data)
self.assertIn("Did this answer your question?".encode(), response.data)
self.assertIn("Yes".encode(), response.data)
self.assertIn("No".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_bricks_with_sub_option_time_to_complete(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
form = {"option": "time-to-complete"}
self.set_flask_session()
response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("How long will it take to complete?".encode(), response.data)
self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data)
self.assertIn("Did this answer your question?".encode(), response.data)
self.assertIn("Yes".encode(), response.data)
self.assertIn("No".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_bricks_with_sub_option_how_long_selected_for(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
form = {"option": "how-long-selected-for"}
self.set_flask_session()
response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("How long will my business be selected for?".encode(), response.data)
self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data)
self.assertIn("Did this answer your question?".encode(), response.data)
self.assertIn("Yes".encode(), response.data)
self.assertIn("No".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_for_voluntary_survey_with_sub_option_penalties(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
form = {"option": "penalties"}
self.set_flask_session()
response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Are there penalties for not completing this survey?".encode(), response.data)
self.assertIn("Did this answer your question?".encode(), response.data)
self.assertNotIn(
"If you do not contact us or complete and return by the deadline, penalties may be incurred "
"resulting in a fine of up to £2,500 (under section 4 of the Statistics of Trade Act 1947, "
"last updated by section 17 of the Criminal Justice Act 1991).".encode(),
response.data,
)
self.assertIn("Yes".encode(), response.data)
self.assertIn("No".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_for_statutory_survey_with_sub_option_penalties(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey_eq
get_business.return_value = business_party
form = {"option": "penalties"}
self.set_flask_session()
response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Are there penalties for not completing this survey?".encode(), response.data)
self.assertIn("Did this answer your question?".encode(), response.data)
self.assertIn(
"If you do not contact us or complete and return by the deadline, penalties may be incurred "
"resulting in a fine of up to £2,500 (under section 4 of the Statistics of Trade Act 1947, "
"last updated by section 17 of the Criminal Justice Act 1991).".encode(),
response.data,
)
self.assertIn("Yes".encode(), response.data)
self.assertIn("No".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_info_bricks_with_sub_option_something_else(self, mock_request, get_survey, get_business):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
form = {"option": "info-something-else"}
self.set_flask_session()
response = self.app.post("/surveys/help/info-about-this-survey", data=form, follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Information about the Monthly Survey of Building Materials Bricks".encode(), response.data)
self.assertIn("https://www.ons.gov.uk/surveys/informationforbusinesses".encode(), response.data)
self.assertIn("Did this answer your question?".encode(), response.data)
self.assertIn("Yes".encode(), response.data)
self.assertIn("No".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_send_message_info_bricks_with_sub_option_exemption_completing_survey(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
self.set_flask_session()
response = self.app.get(
"/surveys/help/info-about-this-survey/" "exemption-completing-survey/send-message",
follow_redirects=True,
)
self.assertEqual(response.status_code, 200)
self.assertIn("Can I be exempt from completing the survey questionnaire?".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_send_message_info_bricks_with_sub_option_why_selected_survey(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
self.set_flask_session()
response = self.app.get(
"/surveys/help/info-about-this-survey/" "why-selected/send-message", follow_redirects=True
)
self.assertEqual(response.status_code, 200)
self.assertIn("How was my business selected?".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_send_message_info_bricks_with_sub_option_time_to_complete(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
self.set_flask_session()
response = self.app.get(
"/surveys/help/info-about-this-survey/" "time-to-complete/send-message",
follow_redirects=True,
)
self.assertEqual(response.status_code, 200)
self.assertIn("How long will it take to complete?".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_send_message_info_bricks_with_sub_option_how_long_selected_for(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
self.set_flask_session()
response = self.app.get(
"/surveys/help/info-about-this-survey/how-long-selected-for/send-message",
follow_redirects=True,
)
self.assertEqual(response.status_code, 200)
self.assertIn("Send a message".encode(), response.data)
self.assertIn("Describe your issue and we will get back to you.".encode(), response.data)
self.assertIn("How long will my business be selected for?".encode(), response.data)
self.assertIn("Enter message".encode(), response.data)
self.assertIn("Send message".encode(), response.data)
self.assertIn("Cancel".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_send_message_info_bricks_with_sub_option_penalties(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
self.set_flask_session()
response = self.app.get("/surveys/help/info-about-this-survey/penalties/send-message", follow_redirects=True)
self.assertEqual(response.status_code, 200)
self.assertIn("Send a message".encode(), response.data)
self.assertIn("Describe your issue and we will get back to you.".encode(), response.data)
self.assertIn("What are the penalties for not completing a survey?".encode(), response.data)
self.assertIn("Enter message".encode(), response.data)
self.assertIn("Send message".encode(), response.data)
self.assertIn("Cancel".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_survey_help_send_message_info_bricks_with_sub_option_info_something_else(
self, mock_request, get_survey, get_business
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
self.set_flask_session()
response = self.app.get(
"/surveys/help/info-about-this-survey/info-something-else/send-message",
follow_redirects=True,
)
self.assertEqual(response.status_code, 200)
self.assertIn("Send a message".encode(), response.data)
self.assertIn("Describe your issue and we will get back to you.".encode(), response.data)
self.assertIn("Information about this survey".encode(), response.data)
self.assertIn("Enter message".encode(), response.data)
self.assertIn("Send message".encode(), response.data)
self.assertIn("Cancel".encode(), response.data)
@requests_mock.mock()
@patch("frontstage.controllers.party_controller.get_respondent_party_by_id")
@patch("frontstage.controllers.party_controller.get_survey_list_details_for_party")
@patch("frontstage.controllers.conversation_controller.send_message")
@patch("frontstage.controllers.party_controller.get_business_by_ru_ref")
@patch("frontstage.controllers.survey_controller.get_survey_by_survey_ref")
def test_create_message_post_success(
self, mock_request, get_survey, get_business, send_message, get_survey_list, get_respondent_party_by_id
):
mock_request.get(url_banner_api, status_code=404)
get_survey.return_value = survey
get_business.return_value = business_party
get_survey_list.return_value = survey_list_todo
get_respondent_party_by_id.return_value = respondent_party
form = {"body": "info-something-else"}
self.set_flask_session()
response = self.app.post(
"/surveys/help/info-about-this-survey/info-something-else/send-message",
data=form,
follow_redirects=True,
)
self.assertEqual(response.status_code, 200)
self.assertIn("Message sent.".encode(), response.data)
| 52.47981 | 120 | 0.717435 | 2,785 | 22,094 | 5.42298 | 0.075404 | 0.080646 | 0.103688 | 0.099053 | 0.90386 | 0.896511 | 0.889889 | 0.882871 | 0.882871 | 0.882871 | 0 | 0.01219 | 0.175704 | 22,094 | 420 | 121 | 52.604762 | 0.816989 | 0.000181 | 0 | 0.748011 | 0 | 0.013263 | 0.309716 | 0.171043 | 0 | 0 | 0 | 0 | 0.281167 | 1 | 0.058355 | false | 0 | 0.013263 | 0 | 0.074271 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
bff941c9a17ec2e4cc0657fbf59116362103afa3 | 6,250 | py | Python | config/sparql_queries.py | ficolo/science-radar | 3929493d93aa3712e4f48e7ab23b659daedb3b1f | [
"Apache-2.0"
] | 1 | 2019-10-03T02:04:12.000Z | 2019-10-03T02:04:12.000Z | config/sparql_queries.py | ficolo/science-radar | 3929493d93aa3712e4f48e7ab23b659daedb3b1f | [
"Apache-2.0"
] | null | null | null | config/sparql_queries.py | ficolo/science-radar | 3929493d93aa3712e4f48e7ab23b659daedb3b1f | [
"Apache-2.0"
] | null | null | null | queries = {
'CO-AUTHORSHIP': """
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX dcterms: <http://purl.org/dc/terms/>
PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
PREFIX foaf: <http://xmlns.com/foaf/0.1/>
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX bibo: <http://purl.org/ontology/bibo/>
PREFIX sio: <http://semanticscience.org/resource/>
SELECT ?node1 ?node2 ?date (COUNT (DISTINCT ?paper) as ?weight){{
{{
SELECT ?paper ?title ?date {{
?paper sio:SIO_001278 ?dataset .
?paper dcterms:title ?title .
?paper dcterms:issued ?date .
FILTER(?date >= "{start_year}-{start_month}-01T00:00:00"^^xsd:dateTime && ?date < "{end_year}-{end_month}-01T00:00:00"^^xsd:dateTime)
}}
}}
?paper bibo:authorList ?authorList .
?authorList rdfs:member ?author1 .
?authorList rdfs:member ?author2 .
?author1 foaf:name ?node1 .
?author2 foaf:name ?node2 .
FILTER(?author1 != ?author2)
}} GROUP BY ?node1 ?node2 ?date
""",
'BURST': """
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX sio: <http://semanticscience.org/resource/>
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX oa: <http://www.w3.org/ns/oa#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX owl: <http://www.w3.org/2002/07/owl#>
PREFIX doco: <http://purl.org/spar/doco/>
PREFIX dcterms: <http://purl.org/dc/terms/>
PREFIX bibo: <http://purl.org/ontology/bibo/>
SELECT ?date ?title ?abstract (GROUP_CONCAT(DISTINCT ?annotation1Label;separator="|") AS ?annotations) {{
?paper sio:SIO_001278 ?dataset .
?paper dcterms:title ?title .
?paper bibo:abstract ?abstract .
?paper dcterms:issued ?date .
?annotation1 a oa:Annotation ;
oa:hasTarget ?paragraph ;
oa:hasBody ?ontoBody1 ;
oa:hasBody ?textualBody1 .
?textualBody1 a oa:TextualBody ;
rdf:value ?annotation1Label .
?paragraph oa:hasSource ?paper .
FILTER(?date >= "{start}-01-01T00:00:00"^^xsd:dateTime && ?date < "{end}-01-01T00:00:00"^^xsd:dateTime)
FILTER(STRSTARTS(STR(?ontoBody1), "{ontology}"))
}} GROUP BY ?date ?title ?abstract
""",
'PAPER_ANNOTATIONS': """
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX sio: <http://semanticscience.org/resource/>
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX oa: <http://www.w3.org/ns/oa#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX luc: <http://www.ontotext.com/owlim/lucene#>
PREFIX owl: <http://www.w3.org/2002/07/owl#>
PREFIX doco: <http://purl.org/spar/doco/>
PREFIX dcterms: <http://purl.org/dc/terms/>
PREFIX bibo: <http://purl.org/ontology/bibo/>
SELECT ?date ?title (GROUP_CONCAT(DISTINCT ?annotation1Label;separator="|") AS ?annotations) {{
{{
SELECT ?paper ?title ?date {{
?paper sio:SIO_001278 ?dataset .
?paper dcterms:title ?title .
?paper dcterms:issued ?date .
FILTER(?date >= "{start_year}-{start_month}-01T00:00:00"^^xsd:dateTime && ?date < "{end_year}-{end_month}-01T00:00:00"^^xsd:dateTime)
}}
}}
?paragraph oa:hasSource ?paper .
?annotation1 a oa:Annotation ;
oa:hasTarget ?paragraph ;
oa:hasBody ?ontoBody1 ;
oa:hasBody ?textualBody1 .
?textualBody1 a oa:TextualBody ;
rdf:value ?annotation1Label .
FILTER(STRSTARTS(STR(?ontoBody1), "http://ncicb.nci.nih.gov/xml/owl/EVS/Thesaurus.owl"))
}} GROUP BY ?date ?title ?abstract
""",
'CO-CITATION': """
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX sio: <http://semanticscience.org/resource/>
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX oa: <http://www.w3.org/ns/oa#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX luc: <http://www.ontotext.com/owlim/lucene#>
PREFIX owl: <http://www.w3.org/2002/07/owl#>
PREFIX doco: <http://purl.org/spar/doco/>
PREFIX dcterms: <http://purl.org/dc/terms/>
PREFIX bibo: <http://purl.org/ontology/bibo/>
SELECT ?title1 ?title2 (COUNT (DISTINCT ?paper) as ?weight) {
{
SELECT ?paper {
?paper sio:SIO_001278 ?dataset .
?paper dcterms:issued ?date .
FILTER(?date >= "{start}-01-01T00:00:00"^^xsd:dateTime && ?date < "{end}-01-01T00:00:00"^^xsd:dateTime)
}
}
?paper bibo:cites ?citedPaper1 .
?paper bibo:cites ?citedPaper2 .
?citedPaper1 dcterms:title ?title1 .
?citedPaper2 dcterms:title ?title2 .
} GROUP BY ?title1 ?title2
""",
'PAPER-CITATIONS': """
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX sio: <http://semanticscience.org/resource/>
PREFIX xsd: <http://www.w3.org/2001/XMLSchema#>
PREFIX oa: <http://www.w3.org/ns/oa#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
PREFIX luc: <http://www.ontotext.com/owlim/lucene#>
PREFIX owl: <http://www.w3.org/2002/07/owl#>
PREFIX doco: <http://purl.org/spar/doco/>
PREFIX dcterms: <http://purl.org/dc/terms/>
PREFIX bibo: <http://purl.org/ontology/bibo/>
SELECT ?date ?title (GROUP_CONCAT (DISTINCT ?citedPaper; separator="|") as ?references) {{
{{
SELECT ?date ?paper ?title {{
?paper sio:SIO_001278 ?dataset .
?paper dcterms:issued ?date .
?paper dcterms:title ?title .
FILTER(?date >= "{start_year}-{start_month}-01T00:00:00"^^xsd:dateTime && ?date < "{end_year}-{end_month}-01T00:00:00"^^xsd:dateTime)
}}
}}
?paper bibo:cites ?citedPaper .
?citedPaper a bibo:AcademicArticle .
}} GROUP BY ?date ?title
"""
} | 44.326241 | 149 | 0.60064 | 797 | 6,250 | 4.683814 | 0.136763 | 0.05813 | 0.067506 | 0.090008 | 0.798286 | 0.785427 | 0.77498 | 0.725422 | 0.712028 | 0.676668 | 0 | 0.066982 | 0.22128 | 6,250 | 141 | 150 | 44.326241 | 0.700021 | 0 | 0 | 0.722628 | 0 | 0.116788 | 0.984962 | 0.102224 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
872ab1779654eb50d33ff0c304ef10703b774223 | 19,028 | py | Python | bayes_layers.py | JurijsNazarovs/bayesian_nn | 936bf55e0a1e620504d5159c100a74493bd16399 | [
"MIT"
] | 6 | 2021-08-25T04:26:41.000Z | 2022-01-24T01:29:40.000Z | bayes_layers.py | JurijsNazarovs/bayesian_nn | 936bf55e0a1e620504d5159c100a74493bd16399 | [
"MIT"
] | 1 | 2021-10-09T07:29:21.000Z | 2021-10-11T16:59:38.000Z | bayes_layers.py | JurijsNazarovs/bayesian_nn | 936bf55e0a1e620504d5159c100a74493bd16399 | [
"MIT"
] | 3 | 2021-09-05T08:58:20.000Z | 2021-12-03T14:54:09.000Z | import torch
from torch import nn
from torch.nn import Parameter
import torch.nn.functional as F
from torch.nn.modules.utils import _single, _pair, _triple
import numpy as np
import math
import vi_posteriors as vip
import importlib
importlib.reload(vip)
minvar = -3
maxvar = -2
# -------------------------------------------------------------------------------
# Main building blocks
# -------------------------------------------------------------------------------
class BayesLocScaleConvBaseBlock(nn.Module):
"""
Necessary to define this block, despite that it looks like _ConvNd
from pyTorch, because we are training posterior parameters to sample weights
and not weights by itself. Thus, self.weights are not presented here.
We assume that approximate posterior of weights can be parametrized by
2 parameters: mu and logsigmasq.
"""
def __init__(self,
in_channels,
out_channels,
kernel_size,
stride,
padding,
dilation,
transposed,
groups,
prior_mu=0,
prior_logsigmasq=0,
is_mixed_prior=False,
prior_p_mixed=1 / 2,
bias=False):
# Default description of module _ConvNd from pytorch
super().__init__()
if in_channels % groups != 0:
raise ValueError('in_channels must be divisible by groups')
if out_channels % groups != 0:
raise ValueError('out_channels must be divisible by groups')
self.in_channels = in_channels
self.out_channels = out_channels
self.kernel_size = kernel_size
self.stride = stride
self.padding = padding
self.dilation = dilation
self.groups = groups
self.bias = bias
# Save parameters to fill prior
self.prior_params = {}
self.prior_params['mu'] = prior_mu
self.prior_params['logsigmasq'] = prior_logsigmasq
self.prior_params['is_mixed'] = is_mixed_prior
self.prior_params['p_mixed'] = prior_p_mixed
# Posterior parameters - trained: mu and logsigmasq
if transposed:
self.post_mu = Parameter(
torch.zeros(
(in_channels, out_channels // groups, *kernel_size)))
self.post_logsigmasq = Parameter(
torch.zeros(
(in_channels, out_channels // groups, *kernel_size)))
else:
self.post_mu = Parameter(
torch.zeros(
(out_channels, in_channels // groups, *kernel_size)))
self.post_logsigmasq = Parameter(
torch.zeros(
(out_channels, in_channels // groups, *kernel_size)))
if self.bias:
self.post_mu_bias = Parameter(
torch.zeros((out_channels, ) + (1, ) * len(kernel_size)))
self.post_logsigmasq_bias = Parameter(
torch.zeros((out_channels, ) + (1, ) * len(kernel_size)))
# Initialize parameters
self.reset_parameters()
self.reset_priors()
def reset_parameters(self):
n = self.in_channels
for k in self.kernel_size:
n *= k
stdv = 1. / math.sqrt(n)
self.post_mu.data.uniform_(-stdv, stdv)
self.post_logsigmasq.data.uniform_(minvar, maxvar) # (0, 1)
if self.bias:
self.post_mu_bias.data.uniform_(-stdv, stdv)
self.post_logsigmasq_bias.data.uniform_(minvar, maxvar) # (0, 1)
def reset_priors(self):
if self.prior_params['is_mixed']:
# Scale mixture prior
logsigmasq1 = -2
logsigmasq2 = -8
u = torch.rand(self.post_logsigmasq.shape,
dtype=self.post_logsigmasq.dtype,
requires_grad=False).to("cpu")
mod_ind = torch.tensor(u <= self.prior_params['p_mixed'],
dtype=self.post_logsigmasq.dtype)
self.prior_mu = Parameter(torch.zeros_like(self.post_mu) +
self.prior_params['mu'],
requires_grad=False)
self.prior_logsigmasq = Parameter(mod_ind * logsigmasq1 +
(1 - mod_ind) * logsigmasq2,
requires_grad=False)
if self.bias:
self.prior_mu_bias = Parameter(
torch.zeros_like(self.post_mu_bias) +
self.prior_params['mu'],
requires_grad=False)
u = torch.rand(self.post_logsigmasq_bias.shape,
dtype=self.post_logsigmasq.dtype).to("cpu")
mod_ind = torch.tensor(u <= self.prior_params['p_mixed'],
dtype=self.post_logsigmasq.dtype)
self.prior_logsigmasq_bias = Parameter(
mod_ind * logsigmasq1 + (1 - mod_ind) * logsigmasq2,
requires_grad=False)
else:
self.prior_mu = Parameter(torch.zeros_like(self.post_mu) +
self.prior_params['mu'],
requires_grad=False)
self.prior_logsigmasq = Parameter(
torch.zeros_like(self.post_logsigmasq) +
self.prior_params['logsigmasq'],
requires_grad=False)
if self.bias:
self.prior_mu_bias = Parameter(
torch.zeros_like(self.post_mu_bias) +
self.prior_params['mu'],
requires_grad=False)
self.prior_logsigmasq_bias = Parameter(
torch.zeros_like(self.post_logsigmasq_bias) +
self.prior_params['logsigmasq'],
requires_grad=False)
def extra_repr(self):
# Displays arguments for a block
s = ('{in_channels}, {out_channels}, kernel_size={kernel_size}, '
'stride={stride}')
if self.padding != (0, ) * len(self.padding):
s += ', padding={padding}'
if self.dilation != (1, ) * len(self.dilation):
s += ', dilation={dilation}'
if self.groups != 1:
s += ', groups={groups}'
s += ', bias = {bias}'
if hasattr(self, "activation"):
s += ', activation={activation}'
if hasattr(self, "approx_post"):
s += ', approx_post={approx_post}'
if hasattr(self, "kl_method"):
s += ', kl_method={kl_method}'
if hasattr(self, "n_mc_iter"):
s += ', n_mc_iter={n_mc_iter}'
if hasattr(self, "compute_kl"):
s += ', compute_kl={compute_kl}'
return s.format(**self.__dict__)
class Conv1d(BayesLocScaleConvBaseBlock):
"""
Bayesian Convolution 1d layer.
Learned parameters are mu and logsigmasq.
"""
def __init__(self,
in_channels,
out_channels,
kernel_size,
stride=1,
padding=0,
groups=1,
dilation=1,
activation=None,
is_mixed_prior=False,
bias=False,
approx_post="Radial",
kl_method="repar",
n_mc_iter=20,
**kwargs):
kernel_size = _single(kernel_size)
stride = _single(stride)
padding = _single(padding)
dilation = _single(dilation)
self.activation = activation
self.approx_post = approx_post
self.kl_method = kl_method
self.n_mc_iter = n_mc_iter
self.compute_kl = True
super().__init__(in_channels,
out_channels,
kernel_size,
stride,
padding,
dilation,
False,
groups,
is_mixed_prior=is_mixed_prior,
bias=bias)
def forward(self, x):
post = eval("vip." + self.approx_post)
output, kl = post.forward(self, x, fun="conv1d")
return output, kl
class Conv2d(BayesLocScaleConvBaseBlock):
"""
Bayesian Convolution 2d layer.
Learned parameters are mu and logsigmasq.
"""
def __init__(self,
in_channels,
out_channels,
kernel_size,
stride=1,
padding=0,
groups=1,
dilation=1,
activation=None,
is_mixed_prior=False,
bias=False,
approx_post="Radial",
kl_method="repar",
n_mc_iter=20,
**kwargs):
kernel_size = _pair(kernel_size)
stride = _pair(stride)
padding = _pair(padding)
dilation = _pair(dilation)
self.activation = activation
self.approx_post = approx_post
self.kl_method = kl_method
self.n_mc_iter = n_mc_iter
self.compute_kl = True
super().__init__(in_channels,
out_channels,
kernel_size,
stride,
padding,
dilation,
False,
groups,
is_mixed_prior=is_mixed_prior,
bias=bias)
def forward(self, x):
post = eval("vip." + self.approx_post)
output, kl = post.forward(self, x, fun="conv2d")
return output, kl
class Conv3d(BayesLocScaleConvBaseBlock):
"""
Bayesian Convolution 3d layer.
Learned parameters are mu and logsigmasq.
"""
def __init__(self,
in_channels,
out_channels,
kernel_size,
stride=1,
padding=0,
groups=1,
dilation=1,
activation=None,
is_mixed_prior=False,
bias=False,
approx_post="Radial",
kl_method="repar",
n_mc_iter=20,
**kwargs):
kernel_size = _triple(kernel_size)
stride = _triple(stride)
padding = _triple(padding)
dilation = _triple(dilation)
self.activation = activation
self.approx_post = approx_post
self.kl_method = kl_method
self.n_mc_iter = n_mc_iter
self.compute_kl = True
super().__init__(in_channels,
out_channels,
kernel_size,
stride,
padding,
dilation,
False,
groups,
is_mixed_prior=is_mixed_prior,
bias=bias)
def forward(self, x):
post = eval("vip." + self.approx_post)
output, kl = post.forward(self, x, fun="conv3d")
return output, kl
# Transpose classes
class ConvTranspose1d(BayesLocScaleConvBaseBlock):
"""
Bayesian Convolution 1d layer.
Learned parameters are mu and logsigmasq.
"""
def __init__(self,
in_channels,
out_channels,
kernel_size,
stride=1,
padding=0,
groups=1,
dilation=1,
activation=None,
is_mixed_prior=False,
bias=False,
approx_post="Radial",
kl_method="repar",
n_mc_iter=20,
**kwargs):
kernel_size = _single(kernel_size)
stride = _single(stride)
padding = _single(padding)
dilation = _single(dilation)
self.activation = activation
self.approx_post = approx_post
self.kl_method = kl_method
self.n_mc_iter = n_mc_iter
self.compute_kl = True
super().__init__(in_channels,
out_channels,
kernel_size,
stride,
padding,
dilation,
True,
groups,
is_mixed_prior=is_mixed_prior,
bias=bias)
def forward(self, x):
post = eval("vip." + self.approx_post)
output, kl = post.forward(self, x, fun="conv_transpose1d")
return output, kl
class ConvTranspose2d(BayesLocScaleConvBaseBlock):
"""
Bayesian Convolution 2d layer.
Learned parameters are mu and logsigmasq.
"""
def __init__(self,
in_channels,
out_channels,
kernel_size,
stride=1,
padding=0,
groups=1,
dilation=1,
activation=None,
is_mixed_prior=False,
bias=False,
approx_post="Radial",
kl_method="repar",
n_mc_iter=20,
**kwargs):
kernel_size = _pair(kernel_size)
stride = _pair(stride)
padding = _pair(padding)
dilation = _pair(dilation)
self.activation = activation
self.approx_post = approx_post
self.kl_method = kl_method
self.n_mc_iter = n_mc_iter
self.compute_kl = True
super().__init__(in_channels,
out_channels,
kernel_size,
stride,
padding,
dilation,
True,
groups,
is_mixed_prior=is_mixed_prior,
bias=bias)
def forward(self, x):
post = eval("vip." + self.approx_post)
output, kl = post.forward(self, x, fun="conv_transpose2d")
return output, kl
class ConvTranspose3d(BayesLocScaleConvBaseBlock):
"""
Bayesian Convolution 3d layer.
Learned parameters are mu and logsigmasq.
"""
def __init__(self,
in_channels,
out_channels,
kernel_size,
stride=1,
padding=0,
groups=1,
dilation=1,
activation=None,
is_mixed_prior=False,
bias=False,
approx_post="Radial",
kl_method="repar",
n_mc_iter=20,
**kwargs):
kernel_size = _triple(kernel_size)
stride = _triple(stride)
padding = _triple(padding)
dilation = _triple(dilation)
self.activation = activation
self.approx_post = approx_post
self.kl_method = kl_method
self.n_mc_iter = n_mc_iter
self.compute_kl = True
super().__init__(in_channels,
out_channels,
kernel_size,
stride,
padding,
dilation,
True,
groups,
is_mixed_prior=is_mixed_prior,
bias=bias)
def forward(self, x):
post = eval("vip." + self.approx_post)
output, kl = post.forward(self, x, fun="conv_transpose3d")
return output, kl
class Linear(nn.Module):
def __init__(self,
in_features,
out_features,
prior_mu=0,
prior_logsigmasq=0,
bias=False,
activation=None,
approx_post="Radial",
kl_method="repar",
n_mc_iter=20,
**kwargs):
super(Linear, self).__init__()
self.in_features = in_features
self.out_features = out_features
self.bias = bias
self.activation = activation
self.approx_post = approx_post
self.kl_method = kl_method
self.n_mc_iter = n_mc_iter
self.compute_kl = True
# Save parameters to fill prior
self.prior_params = {}
self.prior_params['mu'] = prior_mu
self.prior_params['logsigmasq'] = prior_logsigmasq
# Posterior parameters - trained: mu and logsigmasq
self.post_mu = Parameter(torch.zeros(out_features, in_features))
self.post_logsigmasq = Parameter(torch.zeros(out_features,
in_features))
if self.bias:
self.post_mu_bias = Parameter(torch.zeros((out_features)))
self.post_logsigmasq_bias = Parameter(torch.zeros((out_features)))
self.reset_parameters()
self.reset_priors()
def reset_parameters(self):
stdv = 1 / 2 # 1. / math.sqrt(self.post_mu.size(1))
self.post_mu.data.uniform_(-stdv, stdv)
self.post_logsigmasq.data.uniform_(minvar, maxvar) # (0, 1)
if self.bias:
self.post_mu_bias.data.uniform_(-stdv, stdv)
self.post_logsigmasq_bias.data.uniform_(minvar, maxvar) # (0, 1)
def reset_priors(self):
self.prior_mu = Parameter(torch.zeros_like(self.post_mu) +
self.prior_params['mu'],
requires_grad=False)
self.prior_logsigmasq = Parameter(
torch.zeros_like(self.post_logsigmasq) +
self.prior_params['logsigmasq'],
requires_grad=False)
if self.bias:
self.prior_mu_bias = Parameter(
torch.zeros_like(self.post_mu_bias) + self.prior_params['mu'],
requires_grad=False)
self.prior_logsigmasq_bias = Parameter(
torch.zeros_like(self.post_logsigmasq_bias) +
self.prior_params['logsigmasq'],
requires_grad=False)
def extra_repr(self):
# Displays arguments for a block
s = ('{in_features}, {out_features}')
s += ', bias = {bias}'
if hasattr(self, "activation"):
s += ', activation={activation}'
if hasattr(self, "approx_post"):
s += ', approx_post={approx_post}'
if hasattr(self, "kl_method"):
s += ', kl_method={kl_method}'
if hasattr(self, "n_mc_iter"):
s += ', n_mc_iter={n_mc_iter}'
if hasattr(self, "compute_kl"):
s += ', compute_kl={compute_kl}'
return s.format(**self.__dict__)
def forward(self, x):
post = eval("vip." + self.approx_post)
output, kl = post.forward(self, x, fun="linear")
return output, kl
| 33.382456 | 81 | 0.504415 | 1,870 | 19,028 | 4.853476 | 0.094652 | 0.039665 | 0.020824 | 0.037021 | 0.825584 | 0.810269 | 0.771375 | 0.756611 | 0.74978 | 0.748127 | 0 | 0.008347 | 0.401881 | 19,028 | 569 | 82 | 33.441125 | 0.789122 | 0.070212 | 0 | 0.814732 | 0 | 0 | 0.0518 | 0.014376 | 0 | 0 | 0 | 0 | 0 | 1 | 0.046875 | false | 0 | 0.022321 | 0 | 0.107143 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
87488de1bc875a471cd59888fc9c4465319506ed | 139 | py | Python | cnct/client/models/__init__.py | ffaraone/connect-python-openapi-client | d38f51c96634b4fefa1d8d2edce8a2d96fa888f0 | [
"Apache-2.0"
] | null | null | null | cnct/client/models/__init__.py | ffaraone/connect-python-openapi-client | d38f51c96634b4fefa1d8d2edce8a2d96fa888f0 | [
"Apache-2.0"
] | null | null | null | cnct/client/models/__init__.py | ffaraone/connect-python-openapi-client | d38f51c96634b4fefa1d8d2edce8a2d96fa888f0 | [
"Apache-2.0"
] | null | null | null | from cnct.client.models.base import Action, Collection, NS, Resource # noqa
from cnct.client.models.resourceset import ResourceSet # noqa | 69.5 | 76 | 0.805755 | 19 | 139 | 5.894737 | 0.631579 | 0.142857 | 0.25 | 0.357143 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.115108 | 139 | 2 | 77 | 69.5 | 0.910569 | 0.064748 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 7 |
5e498c18659f131e54ab98b2339fc9e124806690 | 135 | py | Python | supabase_py/lib/__init__.py | bariqhibat/supabase-py | b20703d3d4117c911092212a796e53eb2f5286ca | [
"MIT"
] | null | null | null | supabase_py/lib/__init__.py | bariqhibat/supabase-py | b20703d3d4117c911092212a796e53eb2f5286ca | [
"MIT"
] | null | null | null | supabase_py/lib/__init__.py | bariqhibat/supabase-py | b20703d3d4117c911092212a796e53eb2f5286ca | [
"MIT"
] | null | null | null | from supabase_py.lib import auth_client, query_builder, realtime_client
__all__ = ["auth_client", "query_builder", "realtime_client"]
| 33.75 | 71 | 0.807407 | 18 | 135 | 5.444444 | 0.611111 | 0.204082 | 0.306122 | 0.44898 | 0.734694 | 0.734694 | 0 | 0 | 0 | 0 | 0 | 0 | 0.088889 | 135 | 3 | 72 | 45 | 0.796748 | 0 | 0 | 0 | 0 | 0 | 0.288889 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0.5 | 0 | 0.5 | 0 | 1 | 0 | 0 | null | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 7 |
5e4c0a56e98d3890604e94406829c741ecf8e112 | 15,743 | py | Python | src/tests/illumidesk/authenticators/test_setup_course_hook.py | jgwerner/illumidesk.next | a571b086e992e5f4562d3a605ca45b7ee4352def | [
"Apache-2.0"
] | null | null | null | src/tests/illumidesk/authenticators/test_setup_course_hook.py | jgwerner/illumidesk.next | a571b086e992e5f4562d3a605ca45b7ee4352def | [
"Apache-2.0"
] | 2 | 2021-05-25T13:27:12.000Z | 2021-05-25T13:33:34.000Z | src/tests/illumidesk/authenticators/test_setup_course_hook.py | jgwerner/illumidesk.next | a571b086e992e5f4562d3a605ca45b7ee4352def | [
"Apache-2.0"
] | 2 | 2021-01-07T14:30:02.000Z | 2021-04-28T11:49:42.000Z | import json
import os
from jupyterhub.auth import Authenticator
import pytest
from tornado.web import RequestHandler
from tornado.httpclient import AsyncHTTPClient
from unittest.mock import AsyncMock
from unittest.mock import patch
from illumidesk.apis.jupyterhub_api import JupyterHubAPI
from illumidesk.apis.announcement_service import AnnouncementService
from illumidesk.apis.nbgrader_service import NbGraderServiceHelper
from illumidesk.authenticators.authenticator import LTI11Authenticator
from illumidesk.authenticators.authenticator import LTI13Authenticator
from illumidesk.authenticators.authenticator import setup_course_hook
from illumidesk.authenticators.utils import LTIUtils
@pytest.mark.asyncio
async def test_setup_course_hook_is_assigned_to_lti11_authenticator_post_auth_hook():
"""
Does the setup course hook get assigned to the post_auth_hook for the LTI11Authenticator?
"""
authenticator = LTI11Authenticator(post_auth_hook=setup_course_hook)
assert authenticator.post_auth_hook == setup_course_hook
@pytest.mark.asyncio
async def test_setup_course_hook_is_assigned_to_lti13_authenticator_post_auth_hook():
"""
Does the setup course hook get assigned to the post_auth_hook for the LTI13Authenticator?
"""
authenticator = LTI13Authenticator(post_auth_hook=setup_course_hook)
assert authenticator.post_auth_hook == setup_course_hook
@pytest.mark.asyncio()
async def test_setup_course_hook_raises_environment_error_with_missing_org(
monkeypatch, make_auth_state_dict, setup_course_hook_environ, make_mock_request_handler
):
"""
Is an environment error raised when the organization name is missing when calling
the setup_course_hook function?
"""
monkeypatch.setenv('ORGANIZATION_NAME', '')
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict()
with pytest.raises(EnvironmentError):
await local_authenticator.post_auth_hook(local_authenticator, local_handler, local_authentication)
@pytest.mark.asyncio()
async def test_setup_course_hook_calls_normalize_strings(
auth_state_dict,
setup_course_environ,
setup_course_hook_environ,
make_mock_request_handler,
make_http_response,
mock_nbhelper,
):
"""
Does the setup_course_hook return normalized strings for the username and the course_id?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = auth_state_dict
with patch.object(LTIUtils, 'normalize_string', return_value='intro101') as mock_normalize_string:
with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None):
with patch.object(
AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)
):
_ = await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert mock_normalize_string.called
@pytest.mark.asyncio()
async def test_setup_course_hook_raises_json_decode_error_without_client_fetch_response(
monkeypatch,
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_mock_request_handler,
make_http_response,
mock_nbhelper,
):
"""
Does the setup course hook raise a json decode error if the response form the setup course
microservice is null or empty?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict()
with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None):
with patch.object(
AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request, body=None)
):
with pytest.raises(json.JSONDecodeError):
await setup_course_hook(local_authenticator, local_handler, local_authentication)
@pytest.mark.asyncio()
async def test_setup_course_hook_calls_add_student_to_jupyterhub_group_when_role_is_learner(
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_http_response,
make_mock_request_handler,
mock_nbhelper,
):
"""
Is the jupyterhub_api add student to jupyterhub group function called when the user role is
the learner role?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict()
with patch.object(
JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None
) as mock_add_student_to_jupyterhub_group:
with patch.object(AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)):
result = await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert mock_add_student_to_jupyterhub_group.called
@patch('shutil.chown')
@patch('pathlib.Path.mkdir')
@patch('illumidesk.apis.nbgrader_service.Gradebook')
@pytest.mark.asyncio()
async def test_setup_course_hook_calls_add_user_to_nbgrader_gradebook_when_role_is_learner(
mock_mkdir,
mock_chown,
mock_gradebook,
monkeypatch,
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_mock_request_handler,
make_http_response,
):
"""
Is the jupyterhub_api add user to nbgrader gradebook function called when the user role is
the learner role?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict()
with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None):
with patch.object(
NbGraderServiceHelper, 'add_user_to_nbgrader_gradebook', return_value=None
) as mock_add_user_to_nbgrader_gradebook:
with patch.object(
AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)
):
await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert mock_add_user_to_nbgrader_gradebook.called
@pytest.mark.asyncio()
async def test_setup_course_hook_calls_add_instructor_to_jupyterhub_group_when_role_is_instructor(
monkeypatch,
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_mock_request_handler,
make_http_response,
mock_nbhelper,
):
"""
Is the jupyterhub_api add instructor to jupyterhub group function called when the user role is
the instructor role?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict(user_role='Instructor')
with patch.object(
JupyterHubAPI, 'add_instructor_to_jupyterhub_group', return_value=None
) as mock_add_instructor_to_jupyterhub_group:
with patch.object(AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)):
await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert mock_add_instructor_to_jupyterhub_group.called
@pytest.mark.asyncio()
async def test_setup_course_hook_calls_add_instructor_to_jupyterhub_group_when_role_is_TeachingAssistant(
monkeypatch,
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_mock_request_handler,
make_http_response,
mock_nbhelper,
):
"""
Is the jupyterhub_api add instructor to jupyterhub group function called when the user role is
the instructor role?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict(user_role='urn:lti:role:ims/lis/TeachingAssistant')
with patch.object(
JupyterHubAPI, 'add_instructor_to_jupyterhub_group', return_value=None
) as mock_add_instructor_to_jupyterhub_group:
with patch.object(AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)):
await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert mock_add_instructor_to_jupyterhub_group.called
@pytest.mark.asyncio()
async def test_setup_course_hook_does_not_call_add_student_to_jupyterhub_group_when_role_is_instructor(
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_http_response,
make_mock_request_handler,
mock_nbhelper,
):
"""
Is the jupyterhub_api add student to jupyterhub group function called when the user role is
the instructor role?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict(user_role='Instructor')
with patch.object(
JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None
) as mock_add_student_to_jupyterhub_group:
with patch.object(
JupyterHubAPI, 'add_instructor_to_jupyterhub_group', return_value=None
) as mock_add_instructor_to_jupyterhub_group:
with patch.object(
AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)
):
await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert not mock_add_student_to_jupyterhub_group.called
assert mock_add_instructor_to_jupyterhub_group.called
@pytest.mark.asyncio()
async def test_setup_course_hook_does_not_call_add_instructor_to_jupyterhub_group_when_role_is_learner(
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_http_response,
make_mock_request_handler,
mock_nbhelper,
):
"""
Is the jupyterhub_api add instructor to jupyterhub group function not called when the user role is
the learner role?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict()
with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None):
with patch.object(
JupyterHubAPI, 'add_instructor_to_jupyterhub_group', return_value=None
) as mock_add_instructor_to_jupyterhub_group:
with patch.object(
AsyncHTTPClient,
'fetch',
return_value=make_http_response(handler=local_handler.request),
):
await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert not mock_add_instructor_to_jupyterhub_group.called
@pytest.mark.asyncio()
async def test_setup_course_hook_initialize_data_dict(
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_http_response,
make_mock_request_handler,
mock_nbhelper,
):
"""
Is the data dictionary correctly initialized when properly setting the org env-var and and consistent with the
course id value in the auth state?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict()
expected_data = {
'org': 'test-org',
'course_id': 'intro101',
'domain': '127.0.0.1',
}
with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None):
with patch.object(AsyncHTTPClient, 'fetch', return_value=make_http_response(handler=local_handler.request)):
result = await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert expected_data['course_id'] == result['auth_state']['course_id']
assert expected_data['org'] == os.environ.get('ORGANIZATION_NAME')
assert expected_data['domain'] == local_handler.request.host
@pytest.mark.asyncio()
async def test_setup_course_hook_calls_announcement_service_when_is_new_setup(
setup_course_hook_environ,
make_auth_state_dict,
make_http_response,
make_mock_request_handler,
mock_nbhelper,
):
"""
Is the annuncement service called in new setup?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict()
response_args = {'handler': local_handler.request, 'body': {'is_new_setup': True}}
with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None):
with patch.object(
AsyncHTTPClient,
'fetch',
side_effect=[
make_http_response(**response_args),
None,
], # noqa: E231
):
AnnouncementService.add_announcement = AsyncMock(return_value=None)
await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert AnnouncementService.add_announcement.called
@pytest.mark.asyncio()
async def test_is_new_course_initiates_rolling_update(
setup_course_environ,
setup_course_hook_environ,
make_auth_state_dict,
make_http_response,
make_mock_request_handler,
mock_nbhelper,
):
"""
If the course is a new setup does it initiate a rolling update?
"""
local_authenticator = Authenticator(post_auth_hook=setup_course_hook)
local_handler = make_mock_request_handler(RequestHandler, authenticator=local_authenticator)
local_authentication = make_auth_state_dict()
response_args = {'handler': local_handler.request, 'body': {'is_new_setup': True}}
with patch.object(JupyterHubAPI, 'add_student_to_jupyterhub_group', return_value=None):
with patch.object(
AsyncHTTPClient,
'fetch',
side_effect=[
make_http_response(**response_args),
None,
], # noqa: E231
) as mock_client:
AnnouncementService.add_announcement = AsyncMock(return_value=None)
await setup_course_hook(local_authenticator, local_handler, local_authentication)
assert mock_client.called
mock_client.assert_any_call(
'http://setup-course:8000/rolling-update',
headers={'Content-Type': 'application/json'},
body='',
method='POST',
)
mock_client.assert_any_call(
'http://setup-course:8000',
headers={'Content-Type': 'application/json'},
body='{"org": "test-org", "course_id": "intro101", "domain": "127.0.0.1"}',
method='POST',
)
| 40.574742 | 116 | 0.753478 | 1,878 | 15,743 | 5.887114 | 0.087859 | 0.07064 | 0.07869 | 0.047757 | 0.833936 | 0.81042 | 0.794863 | 0.784461 | 0.775868 | 0.763929 | 0 | 0.003949 | 0.179572 | 15,743 | 387 | 117 | 40.679587 | 0.852044 | 0.001334 | 0 | 0.74386 | 0 | 0.003509 | 0.071851 | 0.037422 | 0 | 0 | 0 | 0 | 0.059649 | 1 | 0 | false | 0 | 0.052632 | 0 | 0.052632 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
5ea732ea7b973e5911b7b86625fbf404f700de0e | 35,207 | py | Python | dataset.py | DeVriesMatt/pointMLP-pytorch | e9c09a2038551e83b072353f3fd7e3294463e892 | [
"Apache-2.0"
] | null | null | null | dataset.py | DeVriesMatt/pointMLP-pytorch | e9c09a2038551e83b072353f3fd7e3294463e892 | [
"Apache-2.0"
] | null | null | null | dataset.py | DeVriesMatt/pointMLP-pytorch | e9c09a2038551e83b072353f3fd7e3294463e892 | [
"Apache-2.0"
] | null | null | null | import torch
from torch.utils.data import Dataset
from sklearn.preprocessing import LabelEncoder
import pandas as pd
import os
from pyntcloud import PyntCloud
import numpy as np
import random
def translate_pointcloud(pointcloud):
xyz1 = np.random.uniform(low=2. / 3., high=3. / 2., size=[3])
xyz2 = np.random.uniform(low=-0.2, high=0.2, size=[3])
translated_pointcloud = np.add(np.multiply(pointcloud, xyz1), xyz2).astype('float32')
return translated_pointcloud
def jitter_pointcloud(pointcloud, sigma=1, clip=0.02):
N, C = pointcloud.shape
rotation = np.copy(pointcloud)
rotation += np.clip(sigma * np.random.randn(N, C), -1 * clip, clip)
return rotation
def generate_24_rotations():
res = []
for id in [[0, 1, 2], [1, 2, 0], [2, 0, 1]]:
R = np.identity(3)[:, id].astype(int)
R1= np.asarray([R[:, 0], R[:, 1], R[:, 2]]).T
R2 = np.asarray([-R[:, 0], -R[:, 1], R[:, 2]]).T
R3 = np.asarray([-R[:, 0], R[:, 1], -R[:, 2]]).T
R4 = np.asarray([R[:, 0], -R[:, 1], -R[:, 2]]).T
res += [R1, R2, R3, R4]
for id in [[0, 2, 1], [1, 0, 2], [2, 1, 0]]:
R = np.identity(3)[:, id].astype(int)
R1 = np.asarray([-R[:, 0], -R[:, 1], -R[:, 2]]).T
R2 = np.asarray([-R[:, 0], R[:, 1], R[:, 2]]).T
R3 = np.asarray([R[:, 0], -R[:, 1], R[:, 2]]).T
R4 = np.asarray([R[:, 0], R[:, 1], -R[:, 2]]).T
res += [R1, R2, R3, R4]
return res
def rotate_pointcloud(pointcloud):
# theta = np.random.normal(0, (np.pi**2)/16, 1)[0]
# print(theta)
theta = np.pi * 2 * np.random.choice(24) / 24
rotation_matrix = np.array([[np.cos(theta), -np.sin(theta)], [np.sin(theta), np.cos(theta)]])
rotation = np.copy(pointcloud)
rotation[:, [0, 2]] = pointcloud[:, [0, 2]].dot(rotation_matrix) # random rotation (x,z)
return rotation, theta
def three_d_rotation(pointcloud):
alpha = np.pi * 2 * np.random.choice(24) / 24
beta = np.pi * 2 * np.random.choice(24) / 24
gamma = np.pi * 2 * np.random.choice(24) / 24
rotation_matrix = np.array(
[[np.cos(beta) * np.cos(gamma),
(np.sin(alpha) * np.sin(beta) * np.cos(gamma)) - (np.cos(alpha) * np.cos(gamma)),
(np.cos(alpha) * np.sin(beta) * np.cos(gamma)) + (np.sin(alpha) * np.sin(gamma))],
[np.cos(beta) * np.sin(gamma),
(np.sin(alpha) * np.sin(beta) * np.sin(gamma)) + (np.cos(alpha) * np.cos(gamma)),
(np.cos(alpha) * np.sin(beta) * np.sin(gamma)) - (np.sin(alpha) * np.cos(gamma))],
[-np.sin(beta),
np.sin(alpha) * np.cos(beta),
np.cos(alpha) * np.cos(beta)]]
)
rotation = np.copy(pointcloud)
rotation[:, ] = pointcloud[:, ].dot(rotation_matrix)
return rotation, (alpha, beta, gamma)
class PointCloudDataset(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=False,
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.centring_only = centring_only
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
img_path = os.path.join(
self.img_dir, treatment, self.new_df.loc[idx, "serialNumber"]
)
image = PyntCloud.from_file(img_path + ".ply")
image = torch.tensor(image.points.values)
# TODO: take away after testing
if self.centring_only:
mean = torch.mean(image, 0)
# mean = torch.tensor([[13.4828, 26.5144, 24.4187]])
# std = torch.tensor([[9.2821, 20.4512, 18.9049]])
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (image - mean) / std
# / std
# TODO: _____________________________________________
else:
mean = torch.tensor([[13.4828, 26.5144, 24.4187]])
std = torch.tensor([[9.2821, 20.4512, 18.9049]])
image = (image - mean) / std
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
return image, label, feats
class PointCloudDatasetAll(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=True,
cell_component="cell",
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.centring_only = centring_only
self.cell_component = cell_component
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
if self.cell_component == "cell":
component_path = "stacked_pointcloud"
else:
component_path = "stacked_pointcloud_nucleus"
img_path = os.path.join(
self.img_dir,
plate_num,
component_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
image = PyntCloud.from_file(img_path + ".ply")
image = torch.tensor(image.points.values)
# TODO: take away after testing
if self.centring_only:
mean = torch.mean(image, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (image - mean) / std
else:
mean = torch.tensor([[13.4828, 26.5144, 24.4187]])
std = torch.tensor([[9.2821, 20.4512, 18.9049]])
image = (image - mean) / std
# TODO: _____________________________________________
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, label, serial_number
class PointCloudDatasetAllBoth(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=False,
cell_component="cell",
proximal=1,
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.cell_component = cell_component
self.proximal = proximal
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
& (
(self.annot_df.Treatment == "Nocodazole")
| (self.annot_df.Treatment == "Blebbistatin")
)
& (self.annot_df.Proximal == self.proximal)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
cell_path = "stacked_pointcloud"
nuc_path = "stacked_pointcloud_nucleus"
cell_img_path = os.path.join(
self.img_dir,
plate_num,
cell_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
nuc_img_path = os.path.join(
self.img_dir,
plate_num,
nuc_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
cell = PyntCloud.from_file(cell_img_path + ".ply")
nuc = PyntCloud.from_file(nuc_img_path + ".ply")
cell = torch.tensor(cell.points.values)
nuc = torch.tensor(nuc.points.values)
full = torch.tensor(np.concatenate((cell[:1024], nuc[:1024])))
mean = torch.mean(full, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (full - mean) / std
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
return image, label, feats
class PointCloudDatasetAllBothNotSpec(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=False,
cell_component="cell",
proximal=1,
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.cell_component = cell_component
self.proximal = proximal
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
cell_path = "stacked_pointcloud"
nuc_path = "stacked_pointcloud_nucleus"
cell_img_path = os.path.join(
self.img_dir,
plate_num,
cell_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
nuc_img_path = os.path.join(
self.img_dir,
plate_num,
nuc_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
cell = PyntCloud.from_file(cell_img_path + ".ply")
nuc = PyntCloud.from_file(nuc_img_path + ".ply")
cell = torch.tensor(cell.points.values)
nuc = torch.tensor(nuc.points.values)
full = torch.tensor(np.concatenate((cell[:1024], nuc[:1024])))
mean = torch.mean(full, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (full - mean) / std
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, label, serial_number
class PointCloudDatasetAllBothNotSpec1024(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=True,
cell_component="cell",
proximal=1,
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.cell_component = cell_component
self.proximal = proximal
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
cell_path = "stacked_pointcloud"
nuc_path = "stacked_pointcloud_nucleus"
cell_img_path = os.path.join(
self.img_dir,
plate_num,
cell_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
nuc_img_path = os.path.join(
self.img_dir,
plate_num,
nuc_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
cell = PyntCloud.from_file(cell_img_path + ".ply")
nuc = PyntCloud.from_file(nuc_img_path + ".ply")
cell = torch.tensor(cell.points.values)
nuc = torch.tensor(nuc.points.values)
full = torch.tensor(np.concatenate((cell[:512], nuc[:512])))
mean = torch.mean(full, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (full - mean) / std
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, label, serial_number
class PointCloudDatasetAll1024(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=True,
cell_component="cell",
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.centring_only = centring_only
self.cell_component = cell_component
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
if self.cell_component == "cell":
component_path = "stacked_pointcloud"
else:
component_path = "stacked_pointcloud_nucleus"
img_path = os.path.join(
self.img_dir,
plate_num,
component_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
image = PyntCloud.from_file(img_path + ".ply")
image = torch.tensor(image.points.values)
# TODO: take away after testing
if self.centring_only:
image = image[:1024]
mean = torch.mean(image, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (image - mean) / std
# TODO: _____________________________________________
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, label, serial_number
class PointCloudDatasetAllRotation1024(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=True,
cell_component="cell",
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.centring_only = centring_only
self.cell_component = cell_component
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
if self.cell_component == "cell":
component_path = "stacked_pointcloud"
else:
component_path = "stacked_pointcloud_nucleus"
img_path = os.path.join(
self.img_dir,
plate_num,
component_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
image = PyntCloud.from_file(img_path + ".ply")
image = torch.tensor(image.points.values)
# TODO: take away after testing
image = image[:1024]
mean = torch.mean(image, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (image - mean) / std
rotated_image, angles = three_d_rotation(image)
# TODO: _____________________________________________
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, rotated_image, angles, serial_number
class PointCloudDatasetAllBothNotSpecRotation(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=False,
cell_component="cell",
proximal=1,
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.cell_component = cell_component
self.proximal = proximal
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
cell_path = "stacked_pointcloud"
nuc_path = "stacked_pointcloud_nucleus"
cell_img_path = os.path.join(
self.img_dir,
plate_num,
cell_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
nuc_img_path = os.path.join(
self.img_dir,
plate_num,
nuc_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
cell = PyntCloud.from_file(cell_img_path + ".ply")
nuc = PyntCloud.from_file(nuc_img_path + ".ply")
cell = torch.tensor(cell.points.values)
nuc = torch.tensor(nuc.points.values)
full = torch.tensor(np.concatenate((cell[:1024], nuc[:1024])))
mean = torch.mean(full, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (full - mean) / std
rotated_image, angles = three_d_rotation(image.numpy())
rotated_image = torch.tensor(rotated_image)
angles = torch.tensor(angles)
# TODO: _____________________________________________
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, rotated_image, angles, serial_number
class PointCloudDatasetAllBothNotSpecRotation1024(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=True,
cell_component="cell",
proximal=1,
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.cell_component = cell_component
self.proximal = proximal
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
cell_path = "stacked_pointcloud"
nuc_path = "stacked_pointcloud_nucleus"
cell_img_path = os.path.join(
self.img_dir,
plate_num,
cell_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
nuc_img_path = os.path.join(
self.img_dir,
plate_num,
nuc_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
cell = PyntCloud.from_file(cell_img_path + ".ply")
nuc = PyntCloud.from_file(nuc_img_path + ".ply")
cell = torch.tensor(cell.points.values)
nuc = torch.tensor(nuc.points.values)
full = torch.tensor(np.concatenate((cell[:512], nuc[:512])))
mean = torch.mean(full, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (full - mean) / std
rotated_image, angles = three_d_rotation(image.numpy())
rotated_image = torch.tensor(rotated_image)
angles = torch.tensor(angles)
# TODO: _____________________________________________
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, rotated_image, angles, serial_number
class PointCloudDatasetAllBothNotSpec2DRotation1024(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=True,
cell_component="cell",
proximal=1,
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.cell_component = cell_component
self.proximal = proximal
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
cell_path = "stacked_pointcloud"
nuc_path = "stacked_pointcloud_nucleus"
cell_img_path = os.path.join(
self.img_dir,
plate_num,
cell_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
nuc_img_path = os.path.join(
self.img_dir,
plate_num,
nuc_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
cell = PyntCloud.from_file(cell_img_path + ".ply")
nuc = PyntCloud.from_file(nuc_img_path + ".ply")
cell = torch.tensor(cell.points.values)
nuc = torch.tensor(nuc.points.values)
full = torch.tensor(np.concatenate((cell[:512], nuc[:512])))
mean = torch.mean(full, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (full - mean) / std
rotated_image, angles = rotate_pointcloud(image.numpy())
rotated_image = torch.tensor(rotated_image)
angles = torch.tensor(angles)
# TODO: _____________________________________________
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, rotated_image, angles, serial_number
class PointCloudDatasetAllBothKLDivergranceRotation1024(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=True,
cell_component="cell",
proximal=1,
rotation_matrices=generate_24_rotations(),
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.cell_component = cell_component
self.proximal = proximal
self.rotation_matrices = rotation_matrices
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
cell_path = "stacked_pointcloud"
nuc_path = "stacked_pointcloud_nucleus"
cell_img_path = os.path.join(
self.img_dir,
plate_num,
cell_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
nuc_img_path = os.path.join(
self.img_dir,
plate_num,
nuc_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
cell = PyntCloud.from_file(cell_img_path + ".ply")
nuc = PyntCloud.from_file(nuc_img_path + ".ply")
cell = torch.tensor(cell.points.values)
nuc = torch.tensor(nuc.points.values)
full = torch.tensor(np.concatenate((cell[:512], nuc[:512])))
mean = torch.mean(full, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (full - mean) / std
rotation_matrix = torch.tensor(self.rotation_matrices[random.randrange(0, 24)]).type(torch.FloatTensor)
rotated_image = torch.matmul(image, rotation_matrix)
# TODO: _____________________________________________
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, rotated_image, serial_number
class SimCLR1024Both(Dataset):
def __init__(
self,
annotations_file,
img_dir,
img_size=400,
label_col="Treatment",
transform=None,
target_transform=None,
centring_only=True,
cell_component="cell",
proximal=1,
rotation_matrices=generate_24_rotations(),
):
self.annot_df = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.img_size = img_size
self.label_col = label_col
self.transform = transform
self.target_transform = target_transform
self.cell_component = cell_component
self.proximal = proximal
self.rotation_matrices = rotation_matrices
self.new_df = self.annot_df[
(self.annot_df.xDim <= self.img_size)
& (self.annot_df.yDim <= self.img_size)
& (self.annot_df.zDim <= self.img_size)
].reset_index(drop=True)
# encode label
le = LabelEncoder()
label_col_enc = self.new_df.loc[:, self.label_col]
label_col_enc = le.fit_transform(label_col_enc)
self.new_df["label_col_enc"] = label_col_enc
def __len__(self):
return len(self.new_df)
def __getitem__(self, idx):
# read the image
treatment = self.new_df.loc[idx, "Treatment"]
plate_num = "Plate" + str(self.new_df.loc[idx, "PlateNumber"])
cell_path = "stacked_pointcloud"
nuc_path = "stacked_pointcloud_nucleus"
cell_img_path = os.path.join(
self.img_dir,
plate_num,
cell_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
nuc_img_path = os.path.join(
self.img_dir,
plate_num,
nuc_path,
treatment,
self.new_df.loc[idx, "serialNumber"],
)
cell = PyntCloud.from_file(cell_img_path + ".ply")
nuc = PyntCloud.from_file(nuc_img_path + ".ply")
cell = torch.tensor(cell.points.values)
nuc = torch.tensor(nuc.points.values)
full = torch.tensor(np.concatenate((cell[:512], nuc[:512])))
mean = torch.mean(full, 0)
std = torch.tensor([[20.0, 20.0, 20.0]])
image = (full - mean) / std
rotation_matrix = torch.tensor(
self.rotation_matrices[random.randrange(1, 24)]).type(torch.FloatTensor)
rotated_image = torch.matmul(image, rotation_matrix)
rotated_jitter = jitter_pointcloud(rotated_image)
rotated_jitter_translated = translate_pointcloud(rotated_jitter)
rotation_matrix2 = torch.tensor(
self.rotation_matrices[random.randrange(1, 24)]).type(torch.FloatTensor)
rotated_image2 = torch.matmul(rotated_image, rotation_matrix2)
rotated_jitter2 = jitter_pointcloud(rotated_image2)
rotated_jitter_translated2 = translate_pointcloud(rotated_jitter2)
# TODO: _____________________________________________
# return encoded label as tensor
label = self.new_df.loc[idx, "label_col_enc"]
label = torch.tensor(label)
# return the classical features as torch tensor
feats = self.new_df.iloc[idx, 16:-4]
feats = torch.tensor(feats)
serial_number = self.new_df.loc[idx, "serialNumber"]
return image, rotated_jitter_translated, rotated_jitter_translated2, serial_number
| 32.568918 | 111 | 0.594399 | 4,341 | 35,207 | 4.474315 | 0.04469 | 0.04505 | 0.057921 | 0.047572 | 0.918344 | 0.908871 | 0.908305 | 0.908305 | 0.901097 | 0.896514 | 0 | 0.021246 | 0.291448 | 35,207 | 1,080 | 112 | 32.599074 | 0.757356 | 0.057858 | 0 | 0.877698 | 0 | 0 | 0.051749 | 0.00864 | 0 | 0 | 0 | 0.000926 | 0 | 1 | 0.049161 | false | 0 | 0.009592 | 0.014388 | 0.107914 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
2196527ce4de91e80ddc8c31852b374dd4fbc1a6 | 63 | py | Python | 25/02/object.py | pylangstudy/201707 | c1cc72667f1e0b6e8eef4ee85067d7fa4ca500b6 | [
"CC0-1.0"
] | null | null | null | 25/02/object.py | pylangstudy/201707 | c1cc72667f1e0b6e8eef4ee85067d7fa4ca500b6 | [
"CC0-1.0"
] | 46 | 2017-06-30T22:19:07.000Z | 2017-07-31T22:51:31.000Z | 25/02/object.py | pylangstudy/201707 | c1cc72667f1e0b6e8eef4ee85067d7fa4ca500b6 | [
"CC0-1.0"
] | null | null | null | #class object
print(object)
print(object())
print(dir(object))
| 12.6 | 18 | 0.746032 | 9 | 63 | 5.222222 | 0.444444 | 0.702128 | 0.723404 | 0.93617 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.079365 | 63 | 4 | 19 | 15.75 | 0.810345 | 0.190476 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 0 | null | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 1 | 0 | 7 |
21da568c1c2790d144346e4c561ee722bce0c779 | 9,389 | py | Python | vip/stats/distances.py | VChristiaens/VIP2.7 | 92ce75f1004b4dd1480c3688124225ce8a98aca2 | [
"MIT"
] | null | null | null | vip/stats/distances.py | VChristiaens/VIP2.7 | 92ce75f1004b4dd1480c3688124225ce8a98aca2 | [
"MIT"
] | null | null | null | vip/stats/distances.py | VChristiaens/VIP2.7 | 92ce75f1004b4dd1480c3688124225ce8a98aca2 | [
"MIT"
] | null | null | null | #! /usr/bin/env python
"""
Distance between images.
"""
from __future__ import division
__author__ = 'C. Gomez @ ULg'
__all__ = ['cube_distance',
'cube_distance_to_frame']
import numpy as np
import scipy.stats
from matplotlib import pyplot as plt
from ..var import get_annulus
def cube_distance(array, frame, mode='full', dist='sad', inradius=None,
width=None, plot=True):
""" Computes the distance (or similarity) between frames in a cube, using
one as the reference. Depending on the mode, the whole image can be used,
or just the pixels in a given annulus. The criteria used are:
- the Manhattan distance (SAD or sum of absolute differences),
- the Euclidean distance (square root of the sum of the squared differences),
- the Mean Squared Error,
- the Spearman correlation coefficient,
- the Pearson correlation coefficient,
- the Structural Similarity Index (SSIM).
The SAD, MSE and Ecuclidean criteria are dissimilarity criteria, which
means that 0 is perfect similarity.
The Spearman and Pearson correlation coefficients, vary between -1 and +1
with 0 implying no correlation. Correlations of -1 or +1 imply an exact
linear relationship.
Parameters
----------
array : array_like
Input cube or 3d array.
frame : int
Reference frame in the cube.
mode : {'full','annulus'}, string optional
Whether to use the full frames or a centered annulus.
dist : {'sad','euclidean','mse','pearson','spearman'}, str optional
Which criterion to use.
inradius : None or int, optional
The inner radius when mode is 'annulus'.
width : None or int, optional
The width when mode is 'annulus'.
plot : {True, False}, bool optional
Whether to plot the distances or not.
Returns
-------
lista : array_like
1d array of distances for each frame wrt the reference one.
"""
if not array.ndim ==3:
raise TypeError('The input array is not a cube or 3d array')
lista = []
n = array.shape[0]
if mode=='full':
frame_ref = array[frame]
elif mode=='annulus' and inradius and width:
frame_ref = get_annulus(array[frame], inradius, width, True)
else:
raise TypeError('Mode not recognized or missing parameters')
for i in range(n):
if mode=='full':
framei = array[i]
elif mode=='annulus':
framei = get_annulus(array[i], inradius, width, True)
if dist=='sad':
lista.append(np.sum(abs(frame_ref - framei)))
elif dist=='euclidean':
lista.append(np.sqrt(np.sum((frame_ref - framei)**2)))
elif dist=='mse':
lista.append((np.sum((frame_ref - framei)**2))/len(frame_ref))
elif dist=='pearson':
pears, _ = scipy.stats.pearsonr(frame_ref.ravel(), framei.ravel())
lista.append(pears)
elif dist=='spearman':
spear, _ = scipy.stats.spearmanr(frame_ref.ravel(), framei.ravel())
lista.append(spear)
else:
raise ValueError('Distance not recognized')
lista = np.array(lista)
median_cor = np.median(lista)
mean_cor = np.mean(lista)
if plot:
_, ax = plt.subplots(figsize=(12,6))
x = range(len(lista))
ax.plot(x, lista, '-', color='blue', alpha=0.3)
ax.plot(x, lista, '.', color='blue', alpha=0.5)
ax.vlines(frame, ymin=np.nanmin(lista), ymax=np.nanmax(lista),
colors='green', linestyles='dashed', lw=2, alpha=0.8,
label='Frame '+str(frame))
ax.hlines(median_cor, xmin=-1, xmax=n+1, colors='purple',
linestyles='solid', label='Median value : '+str(median_cor))
ax.hlines(mean_cor, xmin=-1, xmax=n+1, colors='red',
linestyles='solid', label='Mean value : '+str(mean_cor))
plt.xlabel('Frame number')
if dist=='sad':
plt.ylabel('SAD - Manhattan distance')
elif dist=='euclidean':
plt.ylabel('Euclidean distance')
elif dist=='pearson':
plt.ylabel('Pearson correlation coefficient')
elif dist=='spearman':
plt.ylabel('Spearman correlation coefficient')
elif dist=='mse':
plt.ylabel('Mean squared error')
elif dist=='ssim':
plt.ylabel('Structural Similarity Index')
plt.xlim(xmin=-1, xmax=n+1)
plt.minorticks_on()
plt.legend(fancybox=True, framealpha=0.5, fontsize=12, loc='best')
plt.grid(which='both')
return lista
def cube_distance_to_frame(array, frame_ref, mode='full', dist='sad',
inradius=None, width=None, plot=True):
""" Computes the distance (or similarity) between frames in a cube and a
reference image. Depending on the mode, the whole image can be used,
or just the pixels in a given annulus. The criteria used are:
- the Manhattan distance (SAD or sum of absolute differences),
- the Euclidean distance (square root of the sum of the squared differences),
- the Mean Squared Error,
- the Spearman correlation coefficient,
- the Pearson correlation coefficient,
- the Structural Similarity Index (SSIM).
The SAD, MSE and Ecuclidean criteria are dissimilarity criteria, which
means that 0 is perfect similarity.
The Spearman and Pearson correlation coefficients, vary between -1 and +1
with 0 implying no correlation. Correlations of -1 or +1 imply an exact
linear relationship.
The Structural Similarity Index was proposed by Wang et al. 2004.
(http://www.cns.nyu.edu/pub/eero/wang03-reprint.pdf)
SSIM varies between -1 and 1, where 1 means perfect similarity. SSIM
attempts to model the perceived change in the structural information of the
image.
Parameters
----------
array : array_like
Input cube or 3d array.
frame_ref : array_like
Reference image.
mode : {'full','annulus'}, string optional
Whether to use the full frames or a centered annulus.
dist : {'sad','euclidean','mse','pearson','spearman','ssim'}, str optional
Which criterion to use.
inradius : None or int, optional
The inner radius when mode is 'annulus'.
width : None or int, optional
The width when mode is 'annulus'.
plot : {True, False}, bool optional
Whether to plot the distances or not.
Returns
-------
lista : array_like
1d array of distances for each frame wrt the reference one.
"""
if not array.ndim ==3:
raise TypeError('The input array is not a cube or 3d array')
lista = []
n = array.shape[0]
if mode=='full':
frame_ref = frame_ref
elif mode=='annulus' and inradius and width:
frame_ref = get_annulus(frame_ref, inradius, width, True)
else:
raise TypeError('Mode not recognized or missing parameters')
for i in range(n):
if mode=='full':
framei = array[i]
elif mode=='annulus':
framei = get_annulus(array[i], inradius, width, True)
if dist=='sad':
lista.append(np.sum(abs(frame_ref - framei)))
elif dist=='euclidean':
lista.append(np.sqrt(np.sum((frame_ref - framei)**2)))
elif dist=='mse':
lista.append((np.sum((frame_ref - framei)**2))/len(frame_ref))
elif dist=='pearson':
pears, _ = scipy.stats.pearsonr(frame_ref.ravel(), framei.ravel())
lista.append(pears)
elif dist=='spearman':
spear, _ = scipy.stats.spearmanr(frame_ref.ravel(), framei.ravel())
lista.append(spear)
elif dist=='ssim':
lista.append(ssim(frame_ref, framei, win_size=7,
dynamic_range=frame_ref.max() - frame_ref.min()))
else:
raise ValueError('Distance not recognized')
lista = np.array(lista)
median_cor = np.median(lista)
mean_cor = np.mean(lista)
if plot:
_, ax = plt.subplots(figsize=(12,6))
x = range(len(lista))
ax.plot(x, lista, '-', color='blue', alpha=0.3)
ax.plot(x, lista, '.', color='blue', alpha=0.5)
ax.hlines(median_cor, xmin=-1, xmax=n+1, colors='purple',
linestyles='solid', label='Median value : '+str(median_cor))
ax.hlines(mean_cor, xmin=-1, xmax=n+1, colors='red',
linestyles='solid', label='Mean value : '+str(mean_cor))
plt.xlabel('Frame number')
if dist=='sad':
plt.ylabel('SAD - Manhattan distance')
elif dist=='euclidean':
plt.ylabel('Euclidean distance')
elif dist=='pearson':
plt.ylabel('Pearson correlation coefficient')
elif dist=='spearman':
plt.ylabel('Spearman correlation coefficient')
elif dist=='mse':
plt.ylabel('Mean squared error')
elif dist=='ssim':
plt.ylabel('Structural Similarity Index')
plt.xlim(xmin=-1, xmax=n+1)
plt.minorticks_on()
plt.legend(fancybox=True, framealpha=0.5, fontsize=12, loc='best')
plt.grid(which='both')
return lista
| 38.479508 | 81 | 0.604857 | 1,202 | 9,389 | 4.66223 | 0.190516 | 0.032834 | 0.017488 | 0.010707 | 0.8601 | 0.8601 | 0.8601 | 0.8601 | 0.8601 | 0.8601 | 0 | 0.010778 | 0.278624 | 9,389 | 243 | 82 | 38.63786 | 0.816625 | 0.335286 | 0 | 0.845588 | 0 | 0 | 0.155932 | 0.003729 | 0 | 0 | 0 | 0 | 0 | 1 | 0.014706 | false | 0 | 0.036765 | 0 | 0.066176 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
1d0dfb271ba5e998b37d3cae96c4833e8f0b11de | 59 | py | Python | testing.py | ChiragMakkar13/salary | 9e761c497ba72622936db4618bb0b6f9911f0230 | [
"MIT"
] | null | null | null | testing.py | ChiragMakkar13/salary | 9e761c497ba72622936db4618bb0b6f9911f0230 | [
"MIT"
] | null | null | null | testing.py | ChiragMakkar13/salary | 9e761c497ba72622936db4618bb0b6f9911f0230 | [
"MIT"
] | 1 | 2022-02-21T15:31:18.000Z | 2022-02-21T15:31:18.000Z | print("Welcome to bitbucket")
print("Welcome to bitbucket") | 29.5 | 29 | 0.779661 | 8 | 59 | 5.75 | 0.5 | 0.521739 | 0.608696 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.084746 | 59 | 2 | 30 | 29.5 | 0.851852 | 0 | 0 | 1 | 0 | 0 | 0.666667 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 0 | null | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 1 | 0 | 9 |
df062286bbb9b5b0cb2a0e767d6553972b3e1f19 | 259 | py | Python | mhealth/users/helpers.py | theCreedo/indigitous-mental-health-app | e3a66a6b2e7fbd105106d0315f78ea3c93647fe8 | [
"MIT"
] | null | null | null | mhealth/users/helpers.py | theCreedo/indigitous-mental-health-app | e3a66a6b2e7fbd105106d0315f78ea3c93647fe8 | [
"MIT"
] | null | null | null | mhealth/users/helpers.py | theCreedo/indigitous-mental-health-app | e3a66a6b2e7fbd105106d0315f78ea3c93647fe8 | [
"MIT"
] | null | null | null | from werkzeug.security import generate_password_hash
from werkzeug.security import check_password_hash
def hash_pwd(password):
return generate_password_hash(password)
def check_password(hashed, password):
return check_password_hash(hashed, password) | 32.375 | 52 | 0.841699 | 34 | 259 | 6.117647 | 0.352941 | 0.230769 | 0.192308 | 0.25 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0.104247 | 259 | 8 | 53 | 32.375 | 0.896552 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.333333 | false | 1 | 0.333333 | 0.333333 | 1 | 0 | 0 | 0 | 0 | null | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 1 | 1 | 1 | 0 | 0 | 10 |
df151440c4b3219f5b8974dd48f9f511b4d6e7b7 | 192 | py | Python | portfolio_project/views.py | madelinepet/portfolio | c9b1b6318a4ff0018cb55d722d4705057acfe1ee | [
"MIT"
] | null | null | null | portfolio_project/views.py | madelinepet/portfolio | c9b1b6318a4ff0018cb55d722d4705057acfe1ee | [
"MIT"
] | null | null | null | portfolio_project/views.py | madelinepet/portfolio | c9b1b6318a4ff0018cb55d722d4705057acfe1ee | [
"MIT"
] | null | null | null | from django.shortcuts import render
def home_view(request):
return render(request, 'generic/home.html')
def projects_view(request):
return render(request, 'generic/projects.html')
| 19.2 | 51 | 0.755208 | 25 | 192 | 5.72 | 0.52 | 0.153846 | 0.237762 | 0.321678 | 0.517483 | 0.517483 | 0 | 0 | 0 | 0 | 0 | 0 | 0.135417 | 192 | 9 | 52 | 21.333333 | 0.861446 | 0 | 0 | 0 | 0 | 0 | 0.197917 | 0.109375 | 0 | 0 | 0 | 0 | 0 | 1 | 0.4 | false | 0 | 0.2 | 0.4 | 1 | 0 | 1 | 0 | 0 | null | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 1 | 1 | 0 | 0 | 7 |
df586da51d18420f97450825dd0bcc3906f70203 | 163 | py | Python | unitorch/models/detectron2/backbone/__init__.py | fuliucansheng/UniTorch | 47038321593ce4e7eabda555bd58c0cf89482146 | [
"MIT"
] | 2 | 2022-02-05T08:52:00.000Z | 2022-03-27T07:01:34.000Z | unitorch/models/detectron2/backbone/__init__.py | Lixin-Qian/unitorch | 47038321593ce4e7eabda555bd58c0cf89482146 | [
"MIT"
] | null | null | null | unitorch/models/detectron2/backbone/__init__.py | Lixin-Qian/unitorch | 47038321593ce4e7eabda555bd58c0cf89482146 | [
"MIT"
] | 1 | 2022-03-27T07:01:13.000Z | 2022-03-27T07:01:13.000Z | # Copyright (c) FULIUCANSHENG.
# Licensed under the MIT License.
import unitorch.models.detectron2.backbone.swin
import unitorch.models.detectron2.backbone.yolo5
| 27.166667 | 48 | 0.822086 | 20 | 163 | 6.7 | 0.75 | 0.208955 | 0.298507 | 0.447761 | 0.567164 | 0 | 0 | 0 | 0 | 0 | 0 | 0.02027 | 0.092025 | 163 | 5 | 49 | 32.6 | 0.885135 | 0.368098 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | true | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | null | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 0 | 0 | 8 |
10daadd1bc977eee2d42c2b629937df6494c1352 | 17,867 | py | Python | snow/plugins/fix2_configuresnow.py | vPeteWalker/gts21 | 90d827028f5c8f21457de81c0b718f7adbe9a262 | [
"MIT"
] | null | null | null | snow/plugins/fix2_configuresnow.py | vPeteWalker/gts21 | 90d827028f5c8f21457de81c0b718f7adbe9a262 | [
"MIT"
] | null | null | null | snow/plugins/fix2_configuresnow.py | vPeteWalker/gts21 | 90d827028f5c8f21457de81c0b718f7adbe9a262 | [
"MIT"
] | 2 | 2021-04-06T16:59:32.000Z | 2021-05-05T18:50:34.000Z | cat <<EOF > /tmp/ConfigureSNOW.py
from selenium import webdriver
from selenium.webdriver.firefox.options import Options
from selenium.webdriver.common.keys import Keys
from selenium.webdriver.common.action_chains import ActionChains
import os, sys, time
options = Options()
options.headless = True
profile = webdriver.FirefoxProfile()
browser = webdriver.Firefox(options=options,firefox_profile=profile)
# Service Now instance id need to be changes in the url
snow_url = "https://@@{SNOWInstanceURL}@@/nav_to.do?"
snow_username = "admin"
snow_password = "@@{SNOWAdminPassword}@@"
snow_users = "https://@@{SNOWInstanceURL}@@/sys_user_list.do?sysparm_userpref_module%3Dc5aa0fff0a0a0aa7009a39da035ea396%26sysparm_clear_stack%3Dtrue"
snow_groups = "https://@@{SNOWInstanceURL}@@/sys_user_group_list.do?sysparm_userpref_module%3Dc5aa68730a0a0aa70036ced8b58ca05c%26sysparm_clear_stack%3Dtrue"
user_criteria_plugin_url = "https://@@{SNOWInstanceURL}@@/\$allappsmgmt.do?sysparm_redirect%3Dtrue&sysparm_search=User%20Criteria%20Scoped%20API"
glide_validate_property_url = "https://@@{SNOWInstanceURL}@@/sys_properties_list.do?sysparm_query=nameSTARTSWITHglide.sc.guide.tab.validate&sysparm_first_row=1&sysparm_view=&sysparm_choice_query_raw=&sysparm_list_header_search=true"
glide_reset_cascade_url = "https://@@{SNOWInstanceURL}@@/sys_properties_list.do?sysparm_query=nameSTARTSWITHglide.sc.reset_cascade&sysparm_first_row=1&sysparm_view=&sysparm_choice_query_raw=&sysparm_list_header_search=true"
tables_url = "https://@@{SNOWInstanceURL}@@/sys_db_object_list.do?sysparm_query=labelSTARTSWITHCatalog%20Client%20Scripts&sysparm_first_row=1&sysparm_view=&sysparm_choice_query_raw=&sysparm_list_header_search=true"
sys_property = "sys_properties.list"
property_name = "glide.sc.guide.tab.validate"
###Login to the ServiceNow Instance
browser.get(snow_url)
time.sleep(30)
#frame = browser.find_element_by_xpath
browser.switch_to.frame('gsft_main')
time.sleep(5)
browser.find_element_by_id("user_name").send_keys(snow_username)
browser.find_element_by_id("user_password").send_keys(snow_password)
browser.find_element_by_id("sysverb_login").click()
print('Logged in to SNOW')
browser.switch_to_default_content()
#Create user
browser.get(snow_users)
time.sleep(30)
browser.find_element_by_id("sysverb_new").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='sys_user.user_name']").send_keys("midserver")
browser.find_element_by_xpath("//*[@id='sys_user.user_password']").send_keys(snow_password)
browser.find_element_by_xpath("//*[@id='sysverb_insert']").click()
time.sleep(5)
browser.get(snow_users)
time.sleep(10)
browser.find_element_by_link_text("midserver").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='tabs2_list']/span[2]/span/span[2]").click()
time.sleep(5)
browser.find_element_by_id("sysverb_edit_m2m").click()
time.sleep(5)
browser.find_element_by_id("_sys_user_role").send_keys("mid_server")
time.sleep(5)
browser.find_element_by_xpath("//*[text()='mid_server']").click()
time.sleep(5)
browser.find_element_by_id("add_to_collection_button").click()
time.sleep(5)
browser.find_element_by_id("select_0_sysverb_save").click()
time.sleep(5)
browser.find_element_by_id("sysverb_update").click()
time.sleep(15)
print('midserver user account created')
#User Criteria Scoped API Plugin Activation
browser.get(user_criteria_plugin_url)
time.sleep(45)
browser.find_element_by_xpath("//*[@id='appSection']/div/render-plugin/div/div[1]/div[2]/div[2]/div/div/button").click()
time.sleep(10)
browser.switch_to.frame("pluginActivateFrame")
time.sleep(5)
browser.find_element_by_id("ok_button").click()
browser.implicitly_wait(600)
browser.find_element_by_id("sysparm_button_close").click()
time.sleep(2)
print('User Criteria Scoped API plugin activated')
browser.switch_to_default_content()
###glide.sc.guide.tab.validate Property Change
browser.get(glide_validate_property_url)
time.sleep(30)
browser.find_element_by_link_text("glide.sc.guide.tab.validate").click()
time.sleep(10)
browser.find_element_by_id("sys_properties.value").clear()
browser.find_element_by_id("sys_properties.value").send_keys("true")
browser.find_element_by_id("sysverb_update").click()
time.sleep(15)
print('glide.sc.guide.tab.validate updated')
###Table_Access_Script
browser.get(snow_url)
time.sleep(30)
browser.find_element_by_id("filter").send_keys("sys.scripts.do")
browser.find_element_by_id("filter").send_keys(Keys.ENTER)
time.sleep(5)
browser.switch_to.frame('gsft_main')
time.sleep(5)
text_area = browser.find_element_by_xpath("//*[@id='runscript']")
text_area.send_keys(
"var table_array_string='sys_user_has_role,sys_user_group,sys_user_grmember,sys_group_has_role,sc_category,"
"sc_catalog,item_option_new,catalog_ui_policy,catalog_script_client,user_criteria,question,question_choice,sysapproval_approver';")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("var query_string='nameIN'+ table_array_string;")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("var db_objectGR=new GlideRecord('sys_db_object');")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("db_objectGR.addEncodedQuery(query_string);")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("db_objectGR.query();")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("while(db_objectGR.next())")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("{")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("gs.error(db_objectGR.label.getDisplayValue()+\"::\"+db_objectGR.read_access.getDisplayValue()+'-->');")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("db_objectGR.read_access=true;")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("db_objectGR.create_access=true;")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("db_objectGR.update_access=true;")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("db_objectGR.delete_access=false;")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("db_objectGR.update();")
text_area.send_keys(Keys.ENTER)
text_area.send_keys("}")
browser.find_element_by_xpath("/html/body/form/input[2]").click()
time.sleep(15)
print('Table permissions updated')
browser.switch_to_default_content()
##Upload & Install build (v1.0+1.1)
browser.get(snow_url)
time.sleep(30)
browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do")
browser.find_element_by_id("filter").send_keys(Keys.ENTER)
time.sleep(5)
browser.switch_to.frame('gsft_main')
time.sleep(5)
browser.find_element_by_link_text("Import Update Set from XML").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Full Certified Build(v1.0+v1.1).xml")
browser.implicitly_wait(10)
browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click()
browser.implicitly_wait(300)
browser.find_element_by_link_text("Nutanix Calm").click()
browser.implicitly_wait(5)
browser.find_element_by_id("preview_update_set").click()
browser.implicitly_wait(180)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
browser.implicitly_wait(10)
try:
browser.find_element_by_link_text("Skip remote update").click()
browser.implicitly_wait(10)
browser.find_element_by_link_text("Skip remote update").click()
browser.implicitly_wait(10)
browser.find_element_by_link_text("Skip remote update").click()
browser.implicitly_wait(10)
browser.find_element_by_link_text("Skip remote update").click()
browser.implicitly_wait(10)
browser.find_element_by_link_text("Skip remote update").click()
browser.implicitly_wait(10)
except:
print('Skip remote update not there, committing update set')
browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click()
browser.implicitly_wait(600)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
browser.implicitly_wait(10)
print('Calm v1.1 plugin installed')
browser.switch_to_default_content()
######Upload & Install build (v1.2_5)
browser.get(snow_url)
time.sleep(30)
browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do")
browser.find_element_by_id("filter").send_keys(Keys.ENTER)
time.sleep(5)
browser.switch_to.frame('gsft_main')
time.sleep(5)
browser.find_element_by_link_text("Import Update Set from XML").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm V1.2_5.xml")
browser.implicitly_wait(10)
browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click()
browser.implicitly_wait(300)
browser.find_element_by_link_text("Nutanix Calm V1.2_5").click()
browser.implicitly_wait(5)
browser.find_element_by_id("preview_update_set").click()
browser.implicitly_wait(90)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
browser.implicitly_wait(10)
try:
browser.find_element_by_link_text("Skip remote update").click()
browser.implicitly_wait(10)
except:
print('Skip remote update not there')
browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click()
browser.implicitly_wait(600)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
print('Calm v1.2.5 plugin installed')
browser.switch_to_default_content()
######Upload & Install build (v1.3_4)
browser.get(snow_url)
time.sleep(30)
browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do")
browser.find_element_by_id("filter").send_keys(Keys.ENTER)
time.sleep(5)
browser.switch_to.frame('gsft_main')
time.sleep(5)
browser.find_element_by_link_text("Import Update Set from XML").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm V1.3_4.xml")
browser.implicitly_wait(10)
browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click()
browser.implicitly_wait(300)
browser.find_element_by_link_text("Nutanix Calm V1.3_4").click()
browser.implicitly_wait(5)
browser.find_element_by_id("preview_update_set").click()
browser.implicitly_wait(90)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
browser.implicitly_wait(10)
browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click()
browser.implicitly_wait(600)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
print('Calm v1.3.4 plugin installed')
browser.switch_to_default_content()
###glide.sc.reset_cascade Property Change
browser.get(glide_reset_cascade_url)
time.sleep(30)
browser.find_element_by_link_text("glide.sc.reset_cascade").click()
time.sleep(10)
browser.find_element_by_id("sys_properties.value").clear()
browser.find_element_by_id("sys_properties.value").send_keys("true")
browser.find_element_by_id("sysverb_update").click()
time.sleep(15)
print('glide.sc.reset_cascade updated')
###Table permission update
#browser.get(tables_url)
#time.sleep(30)
#browser.find_element_by_link_text("Catalog Client Scripts").click()
#time.sleep(5)
#browser.find_element_by_xpath("//*[@id='tabs2_section']/span[3]/span[1]/span[2]").click()
#time.sleep(5)
#browser.find_element_by_xpath("//*[@id='label.ni.sys_db_object.update_access']").click()
#time.sleep(5)
#browser.find_element_by_id("sysverb_update").click()
#time.sleep(5)
#print('Catalog Client Scripts permissions updated')
######Upload & Install build (v1.4.22)
browser.get(snow_url)
time.sleep(30)
browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do")
browser.find_element_by_id("filter").send_keys(Keys.ENTER)
time.sleep(5)
browser.switch_to.frame('gsft_main')
time.sleep(5)
browser.find_element_by_link_text("Import Update Set from XML").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm V1.4.2.xml")
browser.implicitly_wait(10)
browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click()
browser.implicitly_wait(300)
browser.find_element_by_link_text("Nutanix Calm V1.4.2").click()
browser.implicitly_wait(5)
browser.find_element_by_id("preview_update_set").click()
browser.implicitly_wait(90)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
browser.implicitly_wait(10)
#browser.find_element_by_link_text("Skip remote update").click()
#browser.implicitly_wait(10)
browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click()
browser.implicitly_wait(10)
browser.find_element_by_xpath("//*[@id='ok_button']").click()
browser.implicitly_wait(300)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
print('Calm v1.4.2 plugin installed')
browser.switch_to_default_content()
######Upload & Install build (v1.4.3)
browser.get(snow_url)
time.sleep(30)
browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do")
browser.find_element_by_id("filter").send_keys(Keys.ENTER)
time.sleep(5)
browser.switch_to.frame('gsft_main')
time.sleep(5)
browser.find_element_by_link_text("Import Update Set from XML").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm V1.4.3.xml")
browser.implicitly_wait(10)
browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click()
browser.implicitly_wait(300)
browser.find_element_by_link_text("Nutanix Calm V1.4.3").click()
browser.implicitly_wait(5)
browser.find_element_by_id("preview_update_set").click()
browser.implicitly_wait(90)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
browser.implicitly_wait(10)
browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click()
browser.implicitly_wait(300)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
print('Calm v1.4.3 plugin installed')
browser.switch_to_default_content()
######Upload & Install "Nutanix Calm - User Approval" Workflow
browser.get(snow_url)
time.sleep(30)
browser.find_element_by_id("filter").send_keys("sys_remote_update_set_list.do")
browser.find_element_by_id("filter").send_keys(Keys.ENTER)
time.sleep(5)
browser.switch_to.frame('gsft_main')
time.sleep(5)
browser.find_element_by_link_text("Import Update Set from XML").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='attachFile']").send_keys("/tmp/Nutanix Calm User Approval Workflow 1.0.xml")
browser.implicitly_wait(10)
browser.find_element_by_xpath("/html/body/div[2]/form/div[3]/div[2]/input").click()
browser.implicitly_wait(300)
browser.find_element_by_link_text("User Approval Workflow").click()
browser.implicitly_wait(5)
browser.find_element_by_id("preview_update_set").click()
browser.implicitly_wait(90)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
browser.implicitly_wait(10)
browser.find_element_by_xpath("//*[@id='c38b2cab0a0a0b5000470398d9e60c36']").click()
browser.implicitly_wait(300)
browser.find_element_by_xpath("//*[@id='sysparm_button_close']").click()
print('User Approval Workflow installed')
browser.switch_to_default_content()
#Create Calm Users group
browser.get(snow_groups)
time.sleep(30)
browser.find_element_by_id("sysverb_new").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='sys_user_group.name']").send_keys("Calm Users")
browser.find_element_by_xpath("//*[@id='sysverb_insert']").click()
time.sleep(5)
browser.find_element_by_link_text("Calm Users").click()
time.sleep(5)
browser.find_element_by_id("sysverb_edit_m2m").click()
time.sleep(5)
browser.find_element_by_id("_sys_user_role").send_keys("x_nuta2_nutanix_ca.user")
time.sleep(5)
browser.find_element_by_xpath("//*[text()='x_nuta2_nutanix_ca.user']").click()
time.sleep(5)
browser.find_element_by_id("add_to_collection_button").click()
time.sleep(5)
browser.find_element_by_id("select_0_sysverb_save").click()
time.sleep(5)
browser.find_element_by_id("sysverb_update").click()
time.sleep(5)
print('Calm Users group created')
#Create Nutanix Calm users
browser.get(snow_users)
time.sleep(30)
for x in range(1, 11):
name = "operator{}".format('%02d' % x)
browser.find_element_by_id("sysverb_new").click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='sys_user.user_name']").send_keys(name)
browser.find_element_by_xpath("//*[@id='sys_user.user_password']").send_keys("nutanix/4u")
browser.find_element_by_xpath("//*[@id='sysverb_insert']").click()
time.sleep(5)
browser.find_element_by_link_text(name).click()
time.sleep(5)
browser.find_element_by_xpath("//*[@id='tabs2_list']/span[3]/span/span[2]").click()
time.sleep(3)
browser.find_element_by_xpath("/html/body/div[2]/div[2]/div/div[3]/span/div[2]/div[1]/div/div[1]/button[3]").click()
time.sleep(5)
browser.find_element_by_id("_sys_user_group").send_keys("Calm Users")
time.sleep(5)
browser.find_element_by_xpath("//*[text()='Calm Users']").click()
time.sleep(5)
browser.find_element_by_id("add_to_collection_button").click()
time.sleep(5)
browser.find_element_by_id("select_0_sysverb_save").click()
time.sleep(5)
browser.find_element_by_id("sysverb_update").click()
time.sleep(10)
print('{} user account created'.format(name))
######End
browser.quit()
EOF
wget -P /tmp/ https://raw.githubusercontent.com/nutanix/Calm-Servicenow-Plugin/master/v1.1/Nutanix%20Full%20Certified%20Build\(v1.0%2Bv1.1\).xml
wget -P /tmp/ https://raw.githubusercontent.com/nutanix/Calm-Servicenow-Plugin/master/v1.2/Nutanix%20Calm%20V1.2_5.xml
wget -P /tmp/ https://raw.githubusercontent.com/nutanixworkshops/gts21/master/snow/plugins/Nutanix%20Calm%20V1.3_4.xml
wget -P /tmp/ https://raw.githubusercontent.com/nutanixworkshops/gts21/master/snow/plugins/Nutanix%20Calm%20V1.4.2.xml
wget -P /tmp/ https://raw.githubusercontent.com/nutanixworkshops/gts21/master/snow/plugins/Nutanix%20Calm%20V1.4.3.xml
wget -P /tmp/ https://raw.githubusercontent.com/nutanixworkshops/gts21/master/snow/plugins/Nutanix%20Calm%20User%20Approval%20Workflow%201.0.xml
python3 /tmp/ConfigureSNOW.py
| 44.334988 | 232 | 0.791347 | 2,761 | 17,867 | 4.805143 | 0.093444 | 0.103641 | 0.169594 | 0.188437 | 0.814804 | 0.779151 | 0.76174 | 0.742745 | 0.739956 | 0.716967 | 0 | 0.032987 | 0.046454 | 17,867 | 402 | 233 | 44.445274 | 0.74573 | 0.06095 | 0 | 0.675214 | 0 | 0.017094 | 0.330799 | 0.157318 | 0 | 0 | 0 | 0 | 0 | 0 | null | null | 0.011396 | 0.031339 | null | null | 0.045584 | 0 | 0 | 0 | null | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 |
802ca0574914285367c8339e50cb2681ba6c5519 | 2,523 | py | Python | posix_checkapi/TRACES/POT/ut_lind_net_socketoptions.py | JustinCappos/checkapi | 2508c414869eda3479e1384b1bea65ec1e749d3b | [
"Apache-2.0"
] | null | null | null | posix_checkapi/TRACES/POT/ut_lind_net_socketoptions.py | JustinCappos/checkapi | 2508c414869eda3479e1384b1bea65ec1e749d3b | [
"Apache-2.0"
] | null | null | null | posix_checkapi/TRACES/POT/ut_lind_net_socketoptions.py | JustinCappos/checkapi | 2508c414869eda3479e1384b1bea65ec1e749d3b | [
"Apache-2.0"
] | null | null | null | import lind_test_server
from lind_net_constants import *
SyscallError = lind_test_server.SyscallError
sockfd = lind_test_server.socket_syscall(AF_INET, SOCK_STREAM, 0)
# should work...
lind_test_server.bind_syscall(sockfd,'127.0.0.1',50102)
# let's set some options!!! I'll check for function in different tests...
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==0)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==0)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==0)
# reuseport
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==0)
lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT, 1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==0)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==0)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1)
# linger
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==0)
lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER, 1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==1)
# keepalive
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==0)
lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE, 1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==1)
# let's set some options!!!
lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_SNDBUF, 1000)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_SNDBUF)==1000)
lind_test_server.setsockopt_syscall(sockfd, SOL_SOCKET, SO_RCVBUF, 2000)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_RCVBUF)==2000)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_REUSEPORT)==1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_LINGER)==1)
assert(lind_test_server.getsockopt_syscall(sockfd, SOL_SOCKET, SO_KEEPALIVE)==1)
| 42.05 | 80 | 0.833928 | 378 | 2,523 | 5.171958 | 0.124339 | 0.126854 | 0.221995 | 0.303836 | 0.87468 | 0.856266 | 0.856266 | 0.846036 | 0.846036 | 0.809207 | 0 | 0.021501 | 0.059849 | 2,523 | 59 | 81 | 42.762712 | 0.802698 | 0.056282 | 0 | 0.625 | 0 | 0 | 0.003791 | 0 | 0 | 0 | 0 | 0 | 0.6875 | 1 | 0 | false | 0 | 0.0625 | 0 | 0.0625 | 0 | 0 | 0 | 0 | null | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 9 |
802d2f8aac0966415a060edb9115f53510aa7679 | 22,771 | py | Python | gasp/geometry.py | pgg1610/GASP-python | 84fd6f2e57fb3b90dbccdec77dd2d2453c0fd884 | [
"MIT"
] | 42 | 2016-12-08T03:10:40.000Z | 2022-03-16T09:50:33.000Z | gasp/geometry.py | pgg1610/GASP-python | 84fd6f2e57fb3b90dbccdec77dd2d2453c0fd884 | [
"MIT"
] | 2 | 2017-10-06T02:44:12.000Z | 2019-05-19T07:21:44.000Z | gasp/geometry.py | pgg1610/GASP-python | 84fd6f2e57fb3b90dbccdec77dd2d2453c0fd884 | [
"MIT"
] | 28 | 2016-12-08T02:26:57.000Z | 2022-02-14T06:41:57.000Z | # coding: utf-8
# Copyright (c) Henniggroup.
# Distributed under the terms of the MIT License.
from __future__ import division, unicode_literals, print_function
"""
Geometry module:
This module contains classes to hold geometry-specific data and operations,
including any additional constraints. All geometry classes must implement
pad(), unpad() and get_size() methods.
1. Bulk: Data and operations for 3D bulk structures
2. Sheet: Data and operations for 2D sheet structures
3. Wire: Data and operations for 1D wire structures
4. Cluster: Data and operations for 0D cluster structures
"""
from pymatgen.core.lattice import Lattice
from pymatgen.core.sites import Site
import numpy as np
class Bulk(object):
'''
Contains data and operations specific to bulk structures (so not much...).
'''
def __init__(self):
'''
Makes a Bulk object.
'''
self.shape = 'bulk'
self.max_size = np.inf
self.min_size = -np.inf
self.padding = None
def pad(self, cell, padding='from_geometry'):
'''
Does nothing.
Args:
cell: the Cell to pad
padding: the amount of vacuum padding to add. If set to
'from_geometry', then the value in self.padding is used.
'''
pass
def unpad(self, cell, constraints):
'''
Does nothing.
Args:
cell: the Cell to unpad
constraints: the Constraints of the search
'''
pass
def get_size(self, cell):
'''
Returns 0.
Args:
cell: the Cell whose size to get
'''
return 0
class Sheet(object):
'''
Contains data and operations specific to sheet structures.
'''
def __init__(self, geometry_parameters):
'''
Makes a Sheet, and sets default parameter values if necessary.
Args:
geometry_parameters: a dictionary of parameters
'''
self.shape = 'sheet'
# default values
self.default_max_size = np.inf
self.default_min_size = -np.inf
self.default_padding = 10
# parse the parameters, and set defaults if necessary
# max size
if 'max_size' not in geometry_parameters:
self.max_size = self.default_max_size
elif geometry_parameters['max_size'] in (None, 'default'):
self.max_size = self.default_max_size
else:
self.max_size = geometry_parameters['max_size']
# min size
if 'min_size' not in geometry_parameters:
self.min_size = self.default_min_size
elif geometry_parameters['min_size'] in (None, 'default'):
self.min_size = self.default_min_size
else:
self.min_size = geometry_parameters['min_size']
# padding
if 'padding' not in geometry_parameters:
self.padding = self.default_padding
elif geometry_parameters['padding'] in (None, 'default'):
self.padding = self.default_padding
else:
self.padding = geometry_parameters['padding']
def pad(self, cell, padding='from_geometry'):
'''
Modifies a cell by adding vertical vacuum padding and making the
c-lattice vector normal to the plane of the sheet. The atoms are
shifted to the center of the padded sheet.
Args:
cell: the Cell to pad
padding: the amount of vacuum padding to add (in Angstroms). If not
set, then the value in self.padding is used.
'''
# get the padding amount
if padding == 'from_geometry':
pad_amount = self.padding
else:
pad_amount = padding
# make the padded lattice
cell.rotate_to_principal_directions()
species = cell.species
cartesian_coords = cell.cart_coords
cart_bounds = cell.get_bounding_box(cart_coords=True)
minz = cart_bounds[2][0]
maxz = cart_bounds[2][1]
layer_thickness = maxz - minz
ax = cell.lattice.matrix[0][0]
bx = cell.lattice.matrix[1][0]
by = cell.lattice.matrix[1][1]
padded_lattice = Lattice([[ax, 0.0, 0.0], [bx, by, 0.0],
[0.0, 0.0, layer_thickness + pad_amount]])
# modify the cell to correspond to the padded lattice
cell.modify_lattice(padded_lattice)
site_indices = []
for i in range(len(cell.sites)):
site_indices.append(i)
cell.remove_sites(site_indices)
for i in range(len(cartesian_coords)):
cell.append(species[i], cartesian_coords[i],
coords_are_cartesian=True)
# translate the atoms back into the cell if needed, and shift them to
# the vertical center
cell.translate_atoms_into_cell()
frac_bounds = cell.get_bounding_box(cart_coords=False)
z_center = frac_bounds[2][0] + (frac_bounds[2][1] -
frac_bounds[2][0])/2
translation_vector = [0, 0, 0.5 - z_center]
site_indices = [i for i in range(len(cell.sites))]
cell.translate_sites(site_indices, translation_vector,
frac_coords=True, to_unit_cell=False)
def unpad(self, cell, constraints):
'''
Modifies a cell by removing vertical vacuum padding, leaving only
enough to satisfy the per-species MID constraints, and makes the
c-lattice vector normal to the plane of the sheet (if it isn't
already).
Args:
cell: the Cell to unpad
constraints: the Constraints of the search
'''
# make the unpadded lattice
cell.rotate_to_principal_directions()
species = cell.species
cartesian_coords = cell.cart_coords
layer_thickness = self.get_size(cell)
max_mid = constraints.get_max_mid() + 0.01 # just to be safe...
ax = cell.lattice.matrix[0][0]
bx = cell.lattice.matrix[1][0]
by = cell.lattice.matrix[1][1]
unpadded_lattice = Lattice([[ax, 0.0, 0.0], [bx, by, 0.0],
[0.0, 0.0, layer_thickness + max_mid]])
# modify the cell to correspond to the unpadded lattice
cell.modify_lattice(unpadded_lattice)
site_indices = []
for i in range(len(cell.sites)):
site_indices.append(i)
cell.remove_sites(site_indices)
for i in range(len(cartesian_coords)):
cell.append(species[i], cartesian_coords[i],
coords_are_cartesian=True)
# translate the atoms back into the cell if needed, and shift them to
# the vertical center
cell.translate_atoms_into_cell()
frac_bounds = cell.get_bounding_box(cart_coords=False)
z_center = frac_bounds[2][0] + (frac_bounds[2][1] -
frac_bounds[2][0])/2
translation_vector = [0, 0, 0.5 - z_center]
site_indices = [i for i in range(len(cell.sites))]
cell.translate_sites(site_indices, translation_vector,
frac_coords=True, to_unit_cell=False)
def get_size(self, cell):
'''
Returns the layer thickness of a sheet structure, which is the maximum
vertical distance between atoms in the cell.
Precondition: the cell has already been put into sheet format (c
lattice vector parallel to the z-axis and a and b lattice vectors
in the x-y plane)
Args:
cell: the Cell whose size to get
'''
cart_bounds = cell.get_bounding_box(cart_coords=True)
layer_thickness = cart_bounds[2][1] - cart_bounds[2][0]
return layer_thickness
class Wire(object):
'''
Contains data and operations specific to wire structures.
'''
def __init__(self, geometry_parameters):
'''
Makes a Wire, and sets default parameter values if necessary.
Args:
geometry_parameters: a dictionary of parameters
'''
self.shape = 'wire'
# default values
self.default_max_size = np.inf
self.default_min_size = -np.inf
self.default_padding = 10
# parse the parameters, and set defaults if necessary
# max size
if 'max_size' not in geometry_parameters:
self.max_size = self.default_max_size
elif geometry_parameters['max_size'] in (None, 'default'):
self.max_size = self.default_max_size
else:
self.max_size = geometry_parameters['max_size']
# min size
if 'min_size' not in geometry_parameters:
self.min_size = self.default_min_size
elif geometry_parameters['min_size'] in (None, 'default'):
self.min_size = self.default_min_size
else:
self.min_size = geometry_parameters['min_size']
# padding
if 'padding' not in geometry_parameters:
self.padding = self.default_padding
elif geometry_parameters['padding'] in (None, 'default'):
self.padding = self.default_padding
else:
self.padding = geometry_parameters['padding']
def pad(self, cell, padding='from_geometry'):
'''
Modifies a cell by making the c lattice vector parallel to z-axis, and
adds vacuum padding around the structure in the x and y directions by
replacing a and b lattice vectors with padded vectors along the x and y
axes, respectively. The atoms are shifted to the center of the padded
cell.
Args:
cell: the Cell to pad
padding: the amount of vacuum padding to add (in Angstroms). If not
set, then the value in self.padding is used.
'''
# get the padding amount
if padding == 'from_geometry':
pad_amount = self.padding
else:
pad_amount = padding
# make the padded lattice
cell.rotate_c_parallel_to_z()
species = cell.species
cartesian_coords = cell.cart_coords
cart_bounds = cell.get_bounding_box(cart_coords=True)
x_min = cart_bounds[0][0]
x_max = cart_bounds[0][1]
y_min = cart_bounds[1][0]
y_max = cart_bounds[1][1]
x_extent = x_max - x_min
y_extent = y_max - y_min
cz = cell.lattice.matrix[2][2]
padded_lattice = Lattice([[x_extent + pad_amount, 0, 0],
[0, y_extent + pad_amount, 0], [0, 0, cz]])
# modify the cell to correspond to the padded lattice
cell.modify_lattice(padded_lattice)
site_indices = []
for i in range(len(cell.sites)):
site_indices.append(i)
cell.remove_sites(site_indices)
for i in range(len(cartesian_coords)):
cell.append(species[i], cartesian_coords[i],
coords_are_cartesian=True)
# translate the atoms back into the cell if needed, and shift them to
# the horizontal center
cell.translate_atoms_into_cell()
frac_bounds = cell.get_bounding_box(cart_coords=False)
x_center = frac_bounds[0][0] + (frac_bounds[0][1] -
frac_bounds[0][0])/2
y_center = frac_bounds[1][0] + (frac_bounds[1][1] -
frac_bounds[1][0])/2
translation_vector = [0.5 - x_center, 0.5 - y_center, 0.0]
site_indices = [i for i in range(len(cell.sites))]
cell.translate_sites(site_indices, translation_vector,
frac_coords=True, to_unit_cell=False)
def unpad(self, cell, constraints):
'''
Modifies a cell by removing horizontal vacuum padding around a wire,
leaving only enough to satisfy the per-species MID constraints, and
makes the three lattice vectors lie along the three Cartesian
directions.
Args:
cell: the Cell to unpad
constraints: the Constraints of the search
'''
# make the unpadded lattice
cell.rotate_c_parallel_to_z()
species = cell.species
cartesian_coords = cell.cart_coords
cart_bounds = cell.get_bounding_box(cart_coords=True)
x_min = cart_bounds[0][0]
x_max = cart_bounds[0][1]
y_min = cart_bounds[1][0]
y_max = cart_bounds[1][1]
x_extent = x_max - x_min
y_extent = y_max - y_min
cz = cell.lattice.matrix[2][2]
max_mid = constraints.get_max_mid() + 0.01 # just to be safe...
unpadded_lattice = Lattice([[x_extent + max_mid, 0.0, 0.0],
[0, y_extent + max_mid, 0.0],
[0.0, 0.0, cz]])
# modify the cell to correspond to the unpadded lattice
cell.modify_lattice(unpadded_lattice)
site_indices = []
for i in range(len(cell.sites)):
site_indices.append(i)
cell.remove_sites(site_indices)
for i in range(len(cartesian_coords)):
cell.append(species[i], cartesian_coords[i],
coords_are_cartesian=True)
# translate the atoms back into the cell if needed, and shift them to
# the horizontal center
cell.translate_atoms_into_cell()
frac_bounds = cell.get_bounding_box(cart_coords=False)
x_center = frac_bounds[0][0] + (frac_bounds[0][1] -
frac_bounds[0][0])/2
y_center = frac_bounds[1][0] + (frac_bounds[1][1] -
frac_bounds[1][0])/2
translation_vector = [0.5 - x_center, 0.5 - y_center, 0.0]
site_indices = [i for i in range(len(cell.sites))]
cell.translate_sites(site_indices, translation_vector,
frac_coords=True, to_unit_cell=False)
def get_size(self, cell):
'''
Returns the diameter of a wire structure, defined as the maximum
distance between atoms projected to the x-y plane.
Precondition: the cell has already been put into wire format (c
lattice vector is parallel to z-axis and a and b lattice vectors in
the x-y plane), and all sites are located inside the cell (i.e.,
have fractional coordinates between 0 and 1).
Args:
cell: the Cell whose size to get
'''
max_distance = 0
for site_i in cell.sites:
# make Site versions of each PeriodicSite so that the computed
# distance won't include periodic images
non_periodic_site_i = Site(site_i.species_and_occu,
[site_i.coords[0], site_i.coords[1],
0.0])
for site_j in cell.sites:
non_periodic_site_j = Site(site_j.species_and_occu,
[site_j.coords[0], site_j.coords[1],
0.0])
distance = non_periodic_site_i.distance(non_periodic_site_j)
if distance > max_distance:
max_distance = distance
return max_distance
class Cluster(object):
'''
Contains data and operations specific to clusters.
'''
def __init__(self, geometry_parameters):
'''
Makes a Cluster, and sets default parameter values if necessary.
Args:
geometry_parameters: a dictionary of parameters
'''
self.shape = 'cluster'
# default values
self.default_max_size = np.inf
self.default_min_size = -np.inf
self.default_padding = 10
# parse the parameters, and set defaults if necessary
# max size
if 'max_size' not in geometry_parameters:
self.max_size = self.default_max_size
elif geometry_parameters['max_size'] in (None, 'default'):
self.max_size = self.default_max_size
else:
self.max_size = geometry_parameters['max_size']
# min size
if 'min_size' not in geometry_parameters:
self.min_size = self.default_min_size
elif geometry_parameters['min_size'] in (None, 'default'):
self.min_size = self.default_min_size
else:
self.min_size = geometry_parameters['min_size']
# padding
if 'padding' not in geometry_parameters:
self.padding = self.default_padding
elif geometry_parameters['padding'] in (None, 'default'):
self.padding = self.default_padding
else:
self.padding = geometry_parameters['padding']
def pad(self, cell, padding='from_geometry'):
'''
Modifies a cell by replacing the three lattice vectors with ones along
the three Cartesian directions and adding vacuum padding to each one.
The atoms are shifted to the center of the padded cell.
Args:
cell: the Cell to pad
padding: the amount of vacuum padding to add (in Angstroms). If not
set, then the value in self.padding is used.
'''
# get the padding amount
if padding == 'from_geometry':
pad_amount = self.padding
else:
pad_amount = padding
# make the padded lattice
species = cell.species
cartesian_coords = cell.cart_coords
cart_bounds = cell.get_bounding_box(cart_coords=True)
x_min = cart_bounds[0][0]
x_max = cart_bounds[0][1]
y_min = cart_bounds[1][0]
y_max = cart_bounds[1][1]
z_min = cart_bounds[2][0]
z_max = cart_bounds[2][1]
x_extent = x_max - x_min
y_extent = y_max - y_min
z_extent = z_max - z_min
padded_lattice = Lattice([[x_extent + pad_amount, 0, 0],
[0, y_extent + pad_amount, 0],
[0, 0, z_extent + pad_amount]])
# modify the cell to correspond to the padded lattice
cell.modify_lattice(padded_lattice)
site_indices = []
for i in range(len(cell.sites)):
site_indices.append(i)
cell.remove_sites(site_indices)
for i in range(len(cartesian_coords)):
cell.append(species[i], cartesian_coords[i],
coords_are_cartesian=True)
# translate the atoms back into the cell if needed, and shift them to
# the center
cell.translate_atoms_into_cell()
frac_bounds = cell.get_bounding_box(cart_coords=False)
x_center = frac_bounds[0][0] + (frac_bounds[0][1] -
frac_bounds[0][0])/2
y_center = frac_bounds[1][0] + (frac_bounds[1][1] -
frac_bounds[1][0])/2
z_center = frac_bounds[2][0] + (frac_bounds[2][1] -
frac_bounds[2][0])/2
translation_vector = [0.5 - x_center, 0.5 - y_center, 0.5 - z_center]
site_indices = [i for i in range(len(cell.sites))]
cell.translate_sites(site_indices, translation_vector,
frac_coords=True, to_unit_cell=False)
def unpad(self, cell, constraints):
'''
Modifies a cell by removing vacuum padding in every direction, leaving
only enough to satisfy the per-species MID constraints, and makes the
three lattice vectors lie along the three Cartesian directions.
Args:
cell: the Cell to unpad
constraints: the Constraints of the search
'''
# make the unpadded lattice
species = cell.species
cartesian_coords = cell.cart_coords
cart_bounds = cell.get_bounding_box(cart_coords=True)
x_min = cart_bounds[0][0]
x_max = cart_bounds[0][1]
y_min = cart_bounds[1][0]
y_max = cart_bounds[1][1]
z_min = cart_bounds[2][0]
z_max = cart_bounds[2][1]
x_extent = x_max - x_min
y_extent = y_max - y_min
z_extent = z_max - z_min
max_mid = constraints.get_max_mid() + 0.01 # just to be safe...
unpadded_lattice = Lattice([[x_extent + max_mid, 0.0, 0.0],
[0, y_extent + max_mid, 0.0],
[0.0, 0.0, z_extent + max_mid]])
# modify the cell to correspond to the unpadded lattice
cell.modify_lattice(unpadded_lattice)
site_indices = []
for i in range(len(cell.sites)):
site_indices.append(i)
cell.remove_sites(site_indices)
for i in range(len(cartesian_coords)):
cell.append(species[i], cartesian_coords[i],
coords_are_cartesian=True)
# translate the atoms back into the cell if needed, and shift them to
# the center
cell.translate_atoms_into_cell()
frac_bounds = cell.get_bounding_box(cart_coords=False)
x_center = frac_bounds[0][0] + (frac_bounds[0][1] -
frac_bounds[0][0])/2
y_center = frac_bounds[1][0] + (frac_bounds[1][1] -
frac_bounds[1][0])/2
z_center = frac_bounds[2][0] + (frac_bounds[2][1] -
frac_bounds[2][0])/2
translation_vector = [0.5 - x_center, 0.5 - y_center, 0.5 - z_center]
site_indices = [i for i in range(len(cell.sites))]
cell.translate_sites(site_indices, translation_vector,
frac_coords=True, to_unit_cell=False)
def get_size(self, cell):
'''
Returns the diameter of a cluster structure, defined as the maximum
distance between atoms in the cell.
Precondition: all sites are located inside the cell (i.e., have
fractional coordinates between 0 and 1)
Args:
cell: the Cell whose size to get
'''
max_distance = 0
for site_i in cell.sites:
# make Site versions of each PeriodicSite so that the computed
# distance won't include periodic images
non_periodic_site_i = Site(site_i.species_and_occu, site_i.coords)
for site_j in cell.sites:
non_periodic_site_j = Site(site_j.species_and_occu,
site_j.coords)
distance = non_periodic_site_i.distance(non_periodic_site_j)
if distance > max_distance:
max_distance = distance
return max_distance
| 36.846278 | 79 | 0.58715 | 2,942 | 22,771 | 4.344324 | 0.080557 | 0.010015 | 0.007511 | 0.005633 | 0.892262 | 0.876223 | 0.874266 | 0.850481 | 0.822706 | 0.819419 | 0 | 0.017957 | 0.329893 | 22,771 | 617 | 80 | 36.905997 | 0.819647 | 0.247156 | 0 | 0.885802 | 0 | 0 | 0.024317 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.049383 | false | 0.006173 | 0.012346 | 0 | 0.08642 | 0.003086 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
8042bea693f001a0eec6ce6dadeba016db356440 | 5,770 | py | Python | a10sdk/core/cgnv6/cgnv6_stateful_firewall_alg_sip.py | deepfield/a10sdk-python | bfaa58099f51f085d5e91652d1d1a3fd5c529d5d | [
"Apache-2.0"
] | 16 | 2015-05-20T07:26:30.000Z | 2021-01-23T11:56:57.000Z | a10sdk/core/cgnv6/cgnv6_stateful_firewall_alg_sip.py | deepfield/a10sdk-python | bfaa58099f51f085d5e91652d1d1a3fd5c529d5d | [
"Apache-2.0"
] | 6 | 2015-03-24T22:07:11.000Z | 2017-03-28T21:31:18.000Z | a10sdk/core/cgnv6/cgnv6_stateful_firewall_alg_sip.py | deepfield/a10sdk-python | bfaa58099f51f085d5e91652d1d1a3fd5c529d5d | [
"Apache-2.0"
] | 23 | 2015-03-29T15:43:01.000Z | 2021-06-02T17:12:01.000Z | from a10sdk.common.A10BaseClass import A10BaseClass
class SamplingEnable(A10BaseClass):
"""This class does not support CRUD Operations please use parent.
:param counters1: {"enum": ["all", "stat-request", "stat-response", "method-register", "method-invite", "method-ack", "method-cancel", "method-bye", "method-port-config", "method-prack", "method-subscribe", "method-notify", "method-publish", "method-info", "method-refer", "method-message", "method-update", "method-unknown", "parse-error", "keep-alive", "contact-error", "sdp-error", "rtp-port-no-op", "rtp-rtcp-port-success", "rtp-port-failure", "rtcp-port-failure", "contact-port-no-op", "contact-port-success", "contact-port-failure", "contact-new", "contact-alloc-failure", "contact-eim", "contact-eim-set", "rtp-new", "rtp-alloc-failure", "rtp-eim"], "type": "string", "description": "'all': all; 'stat-request': Request Received; 'stat-response': Response Received; 'method-register': Method REGISTER; 'method-invite': Method INVITE; 'method-ack': Method ACK; 'method-cancel': Method CANCEL; 'method-bye': Method BYE; 'method-port-config': Method OPTIONS; 'method-prack': Method PRACK; 'method-subscribe': Method SUBSCRIBE; 'method-notify': Method NOTIFY; 'method-publish': Method PUBLISH; 'method-info': Method INFO; 'method-refer': Method REFER; 'method-message': Method MESSAGE; 'method-update': Method UPDATE; 'method-unknown': Method Unknown; 'parse-error': Message Parse Error; 'keep-alive': Keep Alive; 'contact-error': Contact Process Error; 'sdp-error': SDP Process Error; 'rtp-port-no-op': RTP Port No Op; 'rtp-rtcp-port-success': RTP RTCP Port Success; 'rtp-port-failure': RTP Port Failure; 'rtcp-port-failure': RTCP Port Failure; 'contact-port-no-op': Contact Port No Op; 'contact-port-success': Contact Port Success; 'contact-port-failure': Contact Port Failure; 'contact-new': Contact Alloc; 'contact-alloc-failure': Contact Alloc Failure; 'contact-eim': Contact EIM; 'contact-eim-set': Contact EIM Set; 'rtp-new': RTP Alloc; 'rtp-alloc-failure': RTP Alloc Failure; 'rtp-eim': RTP EIM; ", "format": "enum"}
:param DeviceProxy: The device proxy for REST operations and session handling. Refer to `common/device_proxy.py`
"""
def __init__(self, **kwargs):
self.ERROR_MSG = ""
self.b_key = "sampling-enable"
self.DeviceProxy = ""
self.counters1 = ""
for keys, value in kwargs.items():
setattr(self,keys, value)
class Sip(A10BaseClass):
""" :param sampling_enable: {"minItems": 1, "items": {"type": "object"}, "uniqueItems": true, "type": "array", "array": [{"properties": {"optional": true, "counters1": {"enum": ["all", "stat-request", "stat-response", "method-register", "method-invite", "method-ack", "method-cancel", "method-bye", "method-port-config", "method-prack", "method-subscribe", "method-notify", "method-publish", "method-info", "method-refer", "method-message", "method-update", "method-unknown", "parse-error", "keep-alive", "contact-error", "sdp-error", "rtp-port-no-op", "rtp-rtcp-port-success", "rtp-port-failure", "rtcp-port-failure", "contact-port-no-op", "contact-port-success", "contact-port-failure", "contact-new", "contact-alloc-failure", "contact-eim", "contact-eim-set", "rtp-new", "rtp-alloc-failure", "rtp-eim"], "type": "string", "description": "'all': all; 'stat-request': Request Received; 'stat-response': Response Received; 'method-register': Method REGISTER; 'method-invite': Method INVITE; 'method-ack': Method ACK; 'method-cancel': Method CANCEL; 'method-bye': Method BYE; 'method-port-config': Method OPTIONS; 'method-prack': Method PRACK; 'method-subscribe': Method SUBSCRIBE; 'method-notify': Method NOTIFY; 'method-publish': Method PUBLISH; 'method-info': Method INFO; 'method-refer': Method REFER; 'method-message': Method MESSAGE; 'method-update': Method UPDATE; 'method-unknown': Method Unknown; 'parse-error': Message Parse Error; 'keep-alive': Keep Alive; 'contact-error': Contact Process Error; 'sdp-error': SDP Process Error; 'rtp-port-no-op': RTP Port No Op; 'rtp-rtcp-port-success': RTP RTCP Port Success; 'rtp-port-failure': RTP Port Failure; 'rtcp-port-failure': RTCP Port Failure; 'contact-port-no-op': Contact Port No Op; 'contact-port-success': Contact Port Success; 'contact-port-failure': Contact Port Failure; 'contact-new': Contact Alloc; 'contact-alloc-failure': Contact Alloc Failure; 'contact-eim': Contact EIM; 'contact-eim-set': Contact EIM Set; 'rtp-new': RTP Alloc; 'rtp-alloc-failure': RTP Alloc Failure; 'rtp-eim': RTP EIM; ", "format": "enum"}}}]}
:param sip_value: {"optional": true, "enum": ["disable"], "type": "string", "description": "'disable': Disable ALG; ", "format": "enum"}
:param uuid: {"description": "uuid of the object", "format": "string", "minLength": 1, "modify-not-allowed": 1, "optional": true, "maxLength": 64, "type": "string"}
:param DeviceProxy: The device proxy for REST operations and session handling. Refer to `common/device_proxy.py`
Class Description::
Configure SIP ALG for NAT stateful firewall (default: enabled).
Class sip supports CRUD Operations and inherits from `common/A10BaseClass`.
This class is the `"PARENT"` class for this module.`
URL for this object::
`https://<Hostname|Ip address>//axapi/v3/cgnv6/stateful-firewall/alg/sip`.
"""
def __init__(self, **kwargs):
self.ERROR_MSG = ""
self.required=[]
self.b_key = "sip"
self.a10_url="/axapi/v3/cgnv6/stateful-firewall/alg/sip"
self.DeviceProxy = ""
self.sampling_enable = []
self.sip_value = ""
self.uuid = ""
for keys, value in kwargs.items():
setattr(self,keys, value)
| 93.064516 | 2,088 | 0.685615 | 751 | 5,770 | 5.242344 | 0.167776 | 0.050292 | 0.024384 | 0.016764 | 0.804928 | 0.804928 | 0.804928 | 0.787656 | 0.770892 | 0.770892 | 0 | 0.005215 | 0.135875 | 5,770 | 61 | 2,089 | 94.590164 | 0.784396 | 0.857539 | 0 | 0.47619 | 0 | 0 | 0.077326 | 0.053735 | 0 | 0 | 0 | 0 | 0 | 1 | 0.095238 | false | 0 | 0.047619 | 0 | 0.238095 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 |
339175e00f73565858eacb3fb6d3de894fac5da4 | 17,785 | py | Python | tests/test_libcloud_extended.py | antoinebourayne/sd2c | c76a0c56d5836caba9e6b90cdf7235516e2dd694 | [
"Apache-2.0"
] | null | null | null | tests/test_libcloud_extended.py | antoinebourayne/sd2c | c76a0c56d5836caba9e6b90cdf7235516e2dd694 | [
"Apache-2.0"
] | null | null | null | tests/test_libcloud_extended.py | antoinebourayne/sd2c | c76a0c56d5836caba9e6b90cdf7235516e2dd694 | [
"Apache-2.0"
] | null | null | null | import logging
import os
from unittest import TestCase
import coloredlogs
from dotenv import dotenv_values, find_dotenv
import unittest.mock
import sd2c as sshcrosscloud
from sd2c import utils
from sd2c.libcloud_extended import ProviderSpecific, SpecificAWS, get_provider_specific_driver
from sd2c.ssh_cross_cloud import SSHCrossCloud
import libcloud.compute.drivers.ec2
import libcloud.compute.drivers.azure_arm
import libcloud.compute.base
import libcloud.compute.providers
def side_effect_create_local_rsa_key_pair_no_key(path):
if path == "a":
return False
if path == "a.pub":
return False
def side_effect_create_local_rsa_key_pair_private(path):
if path == "a":
return True
if path == "a.pub":
return False
class FakeProviderDriver:
def __init__(self):
self.node_id = None
self.volume_name = None
self.list_nodes_empty = False
self.node_state = None
self.terminate = True
self.stop = True
self.start = True
self.list_nodes_called = False
self.ex_stop_node_called = False
self.ex_start_node_called = False
self.destroy_node_called = False
self.destroy_volume_called = False
def list_nodes(self, arg=None):
self.list_nodes_called = True
if self.list_nodes_empty:
return None
else:
return_value = [type('node', (), {})()]
return_value[0].id = self.node_id
return_value[0].state = self.node_state
return return_value
def list_volumes(self, ex_resource_group=None):
return_value = [type('volume', (), {})()]
return_value[0].name = self.volume_name
return return_value
def ex_stop_node(self, node):
self.ex_stop_node_called = True
return self.stop
def ex_start_node(self, node):
self.ex_start_node_called = True
return self.start
def destroy_node(self, node=None, ex_destroy_vhd=False, ex_destroy_nic=False):
self.destroy_node_called = True
return self.terminate
def destroy_volume(self, arg=None):
self.destroy_volume_called = True
return
class TestProviderSpecific(TestCase):
def __init__(self, *args, **kwargs):
super(TestProviderSpecific, self).__init__(*args, **kwargs)
self.command_arg = utils.default_args
@unittest.mock.patch('os.system')
@unittest.mock.patch('os.chmod')
@unittest.mock.patch('os.path.isfile', side_effect=side_effect_create_local_rsa_key_pair_no_key)
def test_create_local_rsa_key_pair_no_key(self, is_path_file, os_chmod, os_system):
ssh = SSHCrossCloud(**self.command_arg)
ssh.ssh_params.rsa_private_key_file_path = "a"
ssh.spe_driver.create_local_rsa_key_pair()
os_system.assert_called_with("ssh-keygen -b 2048 -f a")
os_chmod.assert_called()
is_path_file.assert_called()
@unittest.mock.patch('os.system')
@unittest.mock.patch('os.chmod')
@unittest.mock.patch('os.path.isfile', side_effect=side_effect_create_local_rsa_key_pair_private)
def test_create_local_rsa_key_pair_private_key(self, isfile, chmod, os_system):
ssh = SSHCrossCloud(**self.command_arg)
ssh.ssh_params.rsa_private_key_file_path = "a"
ssh.spe_driver.create_local_rsa_key_pair()
os_system.assert_called_with("ssh-keygen -b 2048 -y -f a > a.pub")
@unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values')
@unittest.mock.patch('os.path.isfile')
def test_create_local_rsa_key_pair_private_key_and_public_key(self, isfile, update_custom_values):
ssh = SSHCrossCloud(**self.command_arg)
isfile.return_value = True
ssh.ssh_params.rsa_private_key_file_path = "a"
assert ssh.spe_driver.create_local_rsa_key_pair() == 0
def test_stop_instance_no_arg(self):
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "terminated"
ssh.spe_driver.driver.stop = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.stop_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.ex_stop_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.stop = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.stop_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.ex_stop_node_called is True
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.stop = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.stop_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.ex_stop_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "terminated"
ssh.spe_driver.driver.stop = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.stop_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.ex_stop_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.stop = False
ssh.ssh_params.sshcrosscloud_instance_id = "a"
with self.assertRaises(Exception):
ssh.spe_driver.stop_instance_no_arg()
def test_start_instance_no_arg(self):
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "not_stopped"
ssh.spe_driver.driver.start = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.start_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.ex_start_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "stopped"
ssh.spe_driver.driver.start = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.start_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.ex_start_node_called is True
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "stopped"
ssh.spe_driver.driver.start = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.start_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.ex_start_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "not_stopped"
ssh.spe_driver.driver.start = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.start_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.ex_start_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "stopped"
ssh.spe_driver.driver.start = False
ssh.ssh_params.sshcrosscloud_instance_id = "a"
with self.assertRaises(Exception):
ssh.spe_driver.start_instance_no_arg()
def test_terminate_instance_no_arg(self):
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.terminate_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.destroy_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.terminate_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.destroy_node_called is True
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.terminate_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.destroy_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
assert ssh.spe_driver.terminate_instance_no_arg() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.destroy_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.terminate = False
ssh.ssh_params.sshcrosscloud_instance_id = "a"
with self.assertRaises(Exception):
ssh.spe_driver.terminate_instance_no_arg()
def test_get_node_any_arg(self):
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = 'foo'
ssh.ssh_params.sshcrosscloud_instance_id = 'foo'
assert ssh.spe_driver.get_node_any_arg().id == 'foo'
ssh.spe_driver.driver.list_nodes_empty = True
with self.assertRaises(Exception):
ssh.spe_driver.get_node_any_arg()
ssh.spe_driver.driver.node_id = 'foo'
with self.assertRaises(Exception):
ssh.spe_driver.get_node_any_arg()
class TestSpecificAWS(TestCase):
command_arg = utils.default_args
@unittest.mock.patch('libcloud.compute.drivers.ec2.EC2NodeDriver')
@unittest.mock.patch.object(sshcrosscloud.libcloud_extended.SpecificAWS, 'get_credentials')
@unittest.mock.patch.object(sshcrosscloud.libcloud_extended.SpecificAWS,
'get_region_from_config_file')
def test_init_specific(self, get_region, get_credentials, ec2_node_driver):
os.environ['IMAGE_NAME'] = 'Fedora'
ssh = SSHCrossCloud(**self.command_arg)
ssh.ssh_params.aws.region = None
get_credentials.return_value = "test1", "test2"
ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params)
assert ssh.spe_driver.init_specific() is not None
assert ssh.ssh_params.instance_user == "fedora"
get_region.assert_called()
get_credentials.assert_called()
ec2_node_driver.assert_called()
@unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values')
@unittest.mock.patch('os.path.isfile')
@unittest.mock.patch('configparser.ConfigParser')
def test_get_region(self, config_parser, is_file, update_custom_values):
is_file.return_value = True
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params)
assert ssh.spe_driver.get_region_from_config_file() is not None
config_parser.assert_called()
@unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values')
@unittest.mock.patch('os.path.isfile')
@unittest.mock.patch('configparser.ConfigParser')
def test_get_credentials(self, config_parser, is_file, update_custom_values):
is_file.return_value = True
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params)
assert len(ssh.spe_driver.get_credentials()) is 2
config_parser.assert_called()
class TestSpecificAzure(TestCase):
def __init__(self, *args, **kwargs):
super(TestSpecificAzure, self).__init__(*args, **kwargs)
self.command_arg = utils.default_args
self.command_arg['provider'] = 'azure'
@unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values')
@unittest.mock.patch('os.path.isfile')
@unittest.mock.patch('configparser.ConfigParser')
def test_get_credentials(self, config_parser, is_file, update_custom_values):
is_file.return_value = True
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params)
assert len(ssh.spe_driver.get_credentials()) is 4
config_parser.assert_called()
def test_terminate_instance(self):
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
ssh.ssh_params.general_name = 'foo'
ssh.spe_driver.driver.volume_name = 'foo'
assert ssh.spe_driver.terminate_instance() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.destroy_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
ssh.ssh_params.general_name = 'foo'
ssh.spe_driver.driver.volume_name = 'foo'
assert ssh.spe_driver.terminate_instance() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.destroy_node_called is True
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
ssh.ssh_params.general_name = 'foo'
ssh.spe_driver.driver.volume_name = 'foo'
assert ssh.spe_driver.terminate_instance() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.destroy_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "b"
ssh.spe_driver.driver.node_state = "terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
ssh.ssh_params.general_name = 'foo'
ssh.spe_driver.driver.volume_name = 'foo'
assert ssh.spe_driver.terminate_instance() is None
assert ssh.spe_driver.driver.list_nodes_called is True
assert ssh.spe_driver.driver.destroy_node_called is False
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.terminate = False
ssh.ssh_params.sshcrosscloud_instance_id = "a"
ssh.ssh_params.general_name = 'foo'
ssh.spe_driver.driver.volume_name = 'foo'
with self.assertRaises(Exception):
ssh.spe_driver.terminate_instance()
ssh.spe_driver.driver = FakeProviderDriver()
ssh.spe_driver.driver.node_id = "a"
ssh.spe_driver.driver.node_state = "not_terminated"
ssh.spe_driver.driver.terminate = True
ssh.ssh_params.sshcrosscloud_instance_id = "a"
ssh.ssh_params.general_name = 'foo'
ssh.spe_driver.driver.volume_name = 'foo'
ssh.spe_driver.terminate_instance()
assert ssh.spe_driver.driver.destroy_volume_called is True
class TestSpecificGPC(TestCase):
def __init__(self, *args, **kwargs):
super(TestSpecificGPC, self).__init__(*args, **kwargs)
self.command_arg = utils.default_args
self.command_arg['provider'] = 'gcp'
@unittest.mock.patch.object(sshcrosscloud.utils.SSHParams, 'update_custom_values')
@unittest.mock.patch('os.path.isfile')
@unittest.mock.patch('configparser.ConfigParser')
def test_get_credentials(self, config_parser, is_file, update_custom_values):
is_file.return_value = True
ssh = SSHCrossCloud(**self.command_arg)
ssh.spe_driver = get_provider_specific_driver(ssh.ssh_params)
assert len(ssh.spe_driver.get_credentials()) is 4
config_parser.assert_called()
| 43.378049 | 102 | 0.705876 | 2,387 | 17,785 | 4.93297 | 0.056975 | 0.083567 | 0.167134 | 0.19414 | 0.85087 | 0.816645 | 0.80518 | 0.785902 | 0.771295 | 0.740892 | 0 | 0.001832 | 0.201912 | 17,785 | 409 | 103 | 43.484108 | 0.827744 | 0 | 0 | 0.694286 | 0 | 0 | 0.049873 | 0.009502 | 0 | 0 | 0 | 0 | 0.211429 | 1 | 0.071429 | false | 0 | 0.04 | 0 | 0.16 | 0 | 0 | 0 | 0 | null | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
33e023b577adb0b77dfce41a990bd9ec1f875db1 | 4,965 | py | Python | tests/hwsim/test_ap_pmf.py | yunus/Hostapd-with-WebID | 2fb6f3a704629990af6dbd3a09d535788ba2d489 | [
"Unlicense"
] | 1 | 2021-11-12T10:52:06.000Z | 2021-11-12T10:52:06.000Z | tests/hwsim/test_ap_pmf.py | arend/hostap | 7c0e1e27575cd273f928fd4820047cba0322718c | [
"Unlicense"
] | null | null | null | tests/hwsim/test_ap_pmf.py | arend/hostap | 7c0e1e27575cd273f928fd4820047cba0322718c | [
"Unlicense"
] | 1 | 2021-11-12T10:52:06.000Z | 2021-11-12T10:52:06.000Z | #!/usr/bin/python
#
# Protected management frames tests
# Copyright (c) 2013, Jouni Malinen <j@w1.fi>
#
# This software may be distributed under the terms of the BSD license.
# See README for more details.
import time
import subprocess
import logging
logger = logging.getLogger(__name__)
import hwsim_utils
import hostapd
from wlantest import Wlantest
def test_ap_pmf_required(dev, apdev):
"""WPA2-PSK AP with PMF required"""
ssid = "test-pmf-required"
wt = Wlantest()
wt.flush()
wt.add_passphrase("12345678")
params = hostapd.wpa2_params(ssid=ssid, passphrase="12345678")
params["wpa_key_mgmt"] = "WPA-PSK-SHA256";
params["ieee80211w"] = "2";
hostapd.add_ap(apdev[0]['ifname'], params)
dev[0].connect(ssid, psk="12345678", ieee80211w="1",
key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2")
hwsim_utils.test_connectivity(dev[0].ifname, apdev[0]['ifname'])
dev[1].connect(ssid, psk="12345678", ieee80211w="2",
key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2")
hwsim_utils.test_connectivity(dev[1].ifname, apdev[0]['ifname'])
hapd = hostapd.Hostapd(apdev[0]['ifname'])
hapd.request("SA_QUERY " + dev[0].p2p_interface_addr())
hapd.request("SA_QUERY " + dev[1].p2p_interface_addr())
wt.require_ap_pmf_mandatory(apdev[0]['bssid'])
wt.require_sta_pmf(apdev[0]['bssid'], dev[0].p2p_interface_addr())
wt.require_sta_pmf_mandatory(apdev[0]['bssid'], dev[1].p2p_interface_addr())
time.sleep(0.1)
if wt.get_sta_counter("valid_saqueryresp_tx", apdev[0]['bssid'],
dev[0].p2p_interface_addr()) < 1:
raise Exception("STA did not reply to SA Query")
if wt.get_sta_counter("valid_saqueryresp_tx", apdev[0]['bssid'],
dev[1].p2p_interface_addr()) < 1:
raise Exception("STA did not reply to SA Query")
def test_ap_pmf_optional(dev, apdev):
"""WPA2-PSK AP with PMF optional"""
ssid = "test-pmf-optional"
wt = Wlantest()
wt.flush()
wt.add_passphrase("12345678")
params = hostapd.wpa2_params(ssid=ssid, passphrase="12345678")
params["wpa_key_mgmt"] = "WPA-PSK";
params["ieee80211w"] = "1";
hostapd.add_ap(apdev[0]['ifname'], params)
dev[0].connect(ssid, psk="12345678", ieee80211w="1",
key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2")
hwsim_utils.test_connectivity(dev[0].ifname, apdev[0]['ifname'])
dev[1].connect(ssid, psk="12345678", ieee80211w="2",
key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2")
hwsim_utils.test_connectivity(dev[1].ifname, apdev[0]['ifname'])
wt.require_ap_pmf_optional(apdev[0]['bssid'])
wt.require_sta_pmf(apdev[0]['bssid'], dev[0].p2p_interface_addr())
wt.require_sta_pmf_mandatory(apdev[0]['bssid'], dev[1].p2p_interface_addr())
def test_ap_pmf_optional_2akm(dev, apdev):
"""WPA2-PSK AP with PMF optional (2 AKMs)"""
ssid = "test-pmf-optional-2akm"
wt = Wlantest()
wt.flush()
wt.add_passphrase("12345678")
params = hostapd.wpa2_params(ssid=ssid, passphrase="12345678")
params["wpa_key_mgmt"] = "WPA-PSK WPA-PSK-SHA256";
params["ieee80211w"] = "1";
hostapd.add_ap(apdev[0]['ifname'], params)
dev[0].connect(ssid, psk="12345678", ieee80211w="1",
key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2")
hwsim_utils.test_connectivity(dev[0].ifname, apdev[0]['ifname'])
dev[1].connect(ssid, psk="12345678", ieee80211w="2",
key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2")
hwsim_utils.test_connectivity(dev[1].ifname, apdev[0]['ifname'])
wt.require_ap_pmf_optional(apdev[0]['bssid'])
wt.require_sta_pmf(apdev[0]['bssid'], dev[0].p2p_interface_addr())
wt.require_sta_key_mgmt(apdev[0]['bssid'], dev[0].p2p_interface_addr(),
"PSK-SHA256")
wt.require_sta_pmf_mandatory(apdev[0]['bssid'], dev[1].p2p_interface_addr())
wt.require_sta_key_mgmt(apdev[0]['bssid'], dev[1].p2p_interface_addr(),
"PSK-SHA256")
def test_ap_pmf_negative(dev, apdev):
"""WPA2-PSK AP without PMF (negative test)"""
ssid = "test-pmf-negative"
wt = Wlantest()
wt.flush()
wt.add_passphrase("12345678")
params = hostapd.wpa2_params(ssid=ssid, passphrase="12345678")
hostapd.add_ap(apdev[0]['ifname'], params)
dev[0].connect(ssid, psk="12345678", ieee80211w="1",
key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2")
hwsim_utils.test_connectivity(dev[0].ifname, apdev[0]['ifname'])
try:
dev[1].connect(ssid, psk="12345678", ieee80211w="2",
key_mgmt="WPA-PSK WPA-PSK-SHA256", proto="WPA2")
hwsim_utils.test_connectivity(dev[1].ifname, apdev[0]['ifname'])
raise Exception("PMF required STA connected to no PMF AP")
except Exception, e:
logger.debug("Ignore expected exception: " + str(e))
wt.require_ap_no_pmf(apdev[0]['bssid'])
| 44.72973 | 80 | 0.652971 | 701 | 4,965 | 4.443652 | 0.156919 | 0.052006 | 0.049438 | 0.045907 | 0.805778 | 0.752809 | 0.752809 | 0.731942 | 0.699839 | 0.699839 | 0 | 0.079872 | 0.180463 | 4,965 | 110 | 81 | 45.136364 | 0.685672 | 0.038671 | 0 | 0.645161 | 0 | 0 | 0.190797 | 0.004775 | 0 | 0 | 0 | 0 | 0 | 0 | null | null | 0.086022 | 0.064516 | null | null | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 7 |
1d736cd5a6542dbd3c419f479be48f9d3af48e6d | 10,140 | py | Python | scrapepy/scrapepy.py | ivekmezovu/ScrapePy | f5ea2d82f7913c73b9ed9d35aa54e29768e50015 | [
"MIT"
] | null | null | null | scrapepy/scrapepy.py | ivekmezovu/ScrapePy | f5ea2d82f7913c73b9ed9d35aa54e29768e50015 | [
"MIT"
] | null | null | null | scrapepy/scrapepy.py | ivekmezovu/ScrapePy | f5ea2d82f7913c73b9ed9d35aa54e29768e50015 | [
"MIT"
] | null | null | null | import urllib.request
from urllib.request import urlretrieve
import urllib
from selenium import webdriver
import os
import time
from selenium.common.exceptions import NoSuchElementException
from tqdm import tqdm
from .settings import Settings
from .util import Util
class ScrapePy:
def __init__(self):
self.options = webdriver.ChromeOptions()
self.options.add_argument('headless')
self.location = Settings.chromedriver_location
self.driver = webdriver.Chrome(chrome_options=self.options, executable_path=self.location)
def scrape_by_username(self, username):
url = 'https://www.instagram.com/{}/'.format(username)
driver = self.driver
driver.get(url)
if driver.find_elements_by_xpath("//div[@class='error-container']"):
print(" >>> Wrong url!")
elif Util.check_if_empty_profile(username) == True:
print(" >>> Account doesn't have any post!")
elif Util.check_if_private_profile(username) == True:
print(" >>> Account is private!")
else:
try:
srcs = []
for i in range(1, 7):
driver.execute_script("window.scrollTo(0, document.body.scrollHeight);")
time.sleep(2)
img = driver.find_elements_by_tag_name('a')
#'//img[@class="FFVAD"]')'''
src = [s.get_attribute('href') for s in img if '.com/p/' in s.get_attribute('href')]
[srcs.append(href) for href in src if href not in srcs]
except NoSuchElementException:
print(" >>> Can't find element!")
src_col = []
i=0
for sr in srcs:
i+=1
driver.get(sr)
print("\n >>> Collecting {}/{}, -> {}".format(i, len(srcs), sr))
time.sleep(2)
try:
a = driver.find_element_by_xpath('//img[@class="FFVAD"]').get_attribute('src')
print(" >>> Collected!")
src_col.append(a)
except NoSuchElementException:
print(" >>> Can't find this element!")
print(src_col)
num = 1
for srces in src_col:
try:
'''for i in tqdm(range(len(src_col))):
urllib.request.urlretrieve(srces, '{}.jpg'.format(num))
print(" >>> Downloaded {}/{})".format(num, len(src_col)))
num+=1'''
for f in tqdm(range(10), ascii=True, desc=" >>> Downloading"):
pass
urllib.request.urlretrieve(srces, '{}.jpg'.format(num))
print(" >>> Downloaded {}/{}".format(num, len(src_col)))
num+=1
except Exception as e:
time.sleep(2)
print(" >>> Can't download, don't know, don't care")
driver.close()
def scrape_by_hashtag(self, hashtag):
url = 'https://www.instagram.com/explore/tags/{}/'.format(hashtag)
driver = self.driver
driver.get(url)
if driver.find_elements_by_xpath("//div[@class='error-container']"):
print(" >>> Wrong url!")
elif Util.check_if_empty_profile(username) == True:
print(" >>> Account doesn't have any post!")
elif Util.check_if_private_profile(username) == True:
print(" >>> Account is private!")
else:
try:
srcs = []
for i in range(1, 7):
driver.execute_script("window.scrollTo(0, document.body.scrollHeight);")
time.sleep(2)
img = driver.find_elements_by_tag_name('a')
#'//img[@class="FFVAD"]')'''
src = [s.get_attribute('href') for s in img if '.com/p/' in s.get_attribute('href')]
[srcs.append(href) for href in src if href not in srcs]
except NoSuchElementException:
print(" >>> Can't find element!")
src_col = []
i=0
for sr in srcs:
i+=1
driver.get(sr)
print("\n >>> Collecting {}/{}, -> {}".format(i, len(srcs), sr))
time.sleep(2)
try:
a = driver.find_element_by_xpath('//img[@class="FFVAD"]').get_attribute('src')
print(" >>> Collected!")
src_col.append(a)
except NoSuchElementException:
print(" >>> Can't find this element!")
print(src_col)
num = 1
for srces in src_col:
try:
for f in tqdm(range(10), ascii=True, desc=" >>> Downloading"):
pass
urllib.request.urlretrieve(srces, '{}.jpg'.format(num))
print(" >>> Downloaded {}/{}".format(num, len(src_col)))
num+=1
except Exception as e:
time.sleep(2)
print(" >>> Can't download, don't know, don't care")
driver.close()
def scrape_videos_by_username(self, username):
url = 'https://www.instagram.com/{}/'.format(username)
driver = self.driver
driver.get(url)
if driver.find_elements_by_xpath("//div[@class='error-container']"):
print(" >>> Wrong url!")
elif Util.check_if_empty_profile(username) == True:
print(" >>> Account doesn't have any post!")
elif Util.check_if_private_profile(username) == True:
print(" >>> Account is private!")
else:
try:
srcs = []
for i in range(1, 7):
driver.execute_script("window.scrollTo(0, document.body.scrollHeight);")
time.sleep(2)
img = driver.find_elements_by_tag_name('a')
#'//img[@class="FFVAD"]')'''
src = [s.get_attribute('href') for s in img if '.com/p/' in s.get_attribute('href')]
[srcs.append(href) for href in src if href not in srcs]
except NoSuchElementException:
print(" >>> Can't find element!")
src_col = []
i=0
for sr in srcs:
i+=1
driver.get(sr)
print("\n >>> Collecting {}/{}, -> {}".format(i, len(srcs), sr))
time.sleep(2)
try:
a = driver.find_element_by_xpath('//video[@class="tWeCl"]').get_attribute('src')
print(" >>> Collected!")
src_col.append(a)
except NoSuchElementException:
print(" >>> Can't find this element!")
print(src_col)
num = 1
for srces in src_col:
try:
for f in tqdm(range(10), ascii=True, desc=" >>> Downloading"):
pass
urllib.request.urlretrieve(srces, '{}.mp4'.format(num))
print(" >>> Downloaded {}/{}".format(num, len(src_col)))
num+=1
except Exception as e:
time.sleep(2)
print(" >>> Can't download, don't know, don't care")
driver.close()
def scrape_videos_by_hashtag(self, hashtag):
url = 'https://www.instagram.com/explore/tags/{}/'.format(hashtag)
driver = self.driver
driver.get(url)
if driver.find_elements_by_xpath("//div[@class='error-container']"):
print(" >>> Wrong url!")
elif Util.check_if_empty_profile(username) == True:
print(" >>> Account doesn't have any post!")
elif Util.check_if_private_profile(username) == True:
print(" >>> Account is private!")
else:
try:
srcs = []
for i in range(1, 7):
driver.execute_script("window.scrollTo(0, document.body.scrollHeight);")
time.sleep(2)
img = driver.find_elements_by_tag_name('a')
#'//img[@class="FFVAD"]')'''
src = [s.get_attribute('href') for s in img if '.com/p/' in s.get_attribute('href')]
[srcs.append(href) for href in src if href not in srcs]
except NoSuchElementException:
print(" >>> Can't find element!")
src_col = []
i=0
for sr in srcs:
i+=1
driver.get(sr)
print("\n >>> Collecting {}/{}, -> {}".format(i, len(srcs), sr))
time.sleep(2)
try:
a = driver.find_element_by_xpath('//video[@class="tWeCl"]').get_attribute('src')
print(" >>> Collected!")
src_col.append(a)
except NoSuchElementException:
print(" >>> Can't find this element!")
print(src_col)
num = 1
for srces in src_col:
try:
for f in tqdm(range(10), ascii=True, desc=" >>> Downloading"):
pass
urllib.request.urlretrieve(srces, '{}.mp4'.format(num))
print(" >>> Downloaded {}/{}".format(num, len(src_col)))
num+=1
except Exception as e:
time.sleep(2)
print(" >>> Can't download, don't know, don't care")
driver.close()
| 31.786834 | 105 | 0.463905 | 1,037 | 10,140 | 4.420444 | 0.124397 | 0.028796 | 0.026178 | 0.019634 | 0.899651 | 0.899651 | 0.899651 | 0.899651 | 0.899651 | 0.899651 | 0 | 0.008493 | 0.407791 | 10,140 | 318 | 106 | 31.886792 | 0.754871 | 0.010651 | 0 | 0.901961 | 0 | 0 | 0.17539 | 0.034192 | 0 | 0 | 0 | 0 | 0 | 1 | 0.02451 | false | 0.019608 | 0.04902 | 0 | 0.078431 | 0.196078 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
d533837f54e91a91c940813975b0b7545b0b0cfd | 68,666 | py | Python | benchmarks/SimResults/_bigLittle_hrrs_spec_tugberk_pinned/cmp_GemsFDTD/power.py | TugberkArkose/MLScheduler | e493b6cbf7b9d29a2c9300d7dd6f0c2f102e4061 | [
"Unlicense"
] | null | null | null | benchmarks/SimResults/_bigLittle_hrrs_spec_tugberk_pinned/cmp_GemsFDTD/power.py | TugberkArkose/MLScheduler | e493b6cbf7b9d29a2c9300d7dd6f0c2f102e4061 | [
"Unlicense"
] | null | null | null | benchmarks/SimResults/_bigLittle_hrrs_spec_tugberk_pinned/cmp_GemsFDTD/power.py | TugberkArkose/MLScheduler | e493b6cbf7b9d29a2c9300d7dd6f0c2f102e4061 | [
"Unlicense"
] | null | null | null | power = {'BUSES': {'Area': 1.33155,
'Bus/Area': 1.33155,
'Bus/Gate Leakage': 0.00662954,
'Bus/Peak Dynamic': 0.0,
'Bus/Runtime Dynamic': 0.0,
'Bus/Subthreshold Leakage': 0.0691322,
'Bus/Subthreshold Leakage with power gating': 0.0259246,
'Gate Leakage': 0.00662954,
'Peak Dynamic': 0.0,
'Runtime Dynamic': 0.0,
'Subthreshold Leakage': 0.0691322,
'Subthreshold Leakage with power gating': 0.0259246},
'Core': [{'Area': 32.6082,
'Execution Unit/Area': 8.2042,
'Execution Unit/Complex ALUs/Area': 0.235435,
'Execution Unit/Complex ALUs/Gate Leakage': 0.0132646,
'Execution Unit/Complex ALUs/Peak Dynamic': 0.0704087,
'Execution Unit/Complex ALUs/Runtime Dynamic': 0.25799,
'Execution Unit/Complex ALUs/Subthreshold Leakage': 0.20111,
'Execution Unit/Complex ALUs/Subthreshold Leakage with power gating': 0.0754163,
'Execution Unit/Floating Point Units/Area': 4.6585,
'Execution Unit/Floating Point Units/Gate Leakage': 0.0656156,
'Execution Unit/Floating Point Units/Peak Dynamic': 0.397975,
'Execution Unit/Floating Point Units/Runtime Dynamic': 0.304033,
'Execution Unit/Floating Point Units/Subthreshold Leakage': 0.994829,
'Execution Unit/Floating Point Units/Subthreshold Leakage with power gating': 0.373061,
'Execution Unit/Gate Leakage': 0.122718,
'Execution Unit/Instruction Scheduler/Area': 2.17927,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Area': 0.328073,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Gate Leakage': 0.00115349,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Peak Dynamic': 1.20978,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Runtime Dynamic': 0.188804,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage': 0.017004,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage with power gating': 0.00962066,
'Execution Unit/Instruction Scheduler/Gate Leakage': 0.00730101,
'Execution Unit/Instruction Scheduler/Instruction Window/Area': 1.00996,
'Execution Unit/Instruction Scheduler/Instruction Window/Gate Leakage': 0.00529112,
'Execution Unit/Instruction Scheduler/Instruction Window/Peak Dynamic': 2.07911,
'Execution Unit/Instruction Scheduler/Instruction Window/Runtime Dynamic': 0.32694,
'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage': 0.0800117,
'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage with power gating': 0.0455351,
'Execution Unit/Instruction Scheduler/Peak Dynamic': 4.84781,
'Execution Unit/Instruction Scheduler/ROB/Area': 0.841232,
'Execution Unit/Instruction Scheduler/ROB/Gate Leakage': 0.000856399,
'Execution Unit/Instruction Scheduler/ROB/Peak Dynamic': 1.55892,
'Execution Unit/Instruction Scheduler/ROB/Runtime Dynamic': 0.187509,
'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage': 0.0178624,
'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage with power gating': 0.00897339,
'Execution Unit/Instruction Scheduler/Runtime Dynamic': 0.703254,
'Execution Unit/Instruction Scheduler/Subthreshold Leakage': 0.114878,
'Execution Unit/Instruction Scheduler/Subthreshold Leakage with power gating': 0.0641291,
'Execution Unit/Integer ALUs/Area': 0.47087,
'Execution Unit/Integer ALUs/Gate Leakage': 0.0265291,
'Execution Unit/Integer ALUs/Peak Dynamic': 0.125609,
'Execution Unit/Integer ALUs/Runtime Dynamic': 0.101344,
'Execution Unit/Integer ALUs/Subthreshold Leakage': 0.40222,
'Execution Unit/Integer ALUs/Subthreshold Leakage with power gating': 0.150833,
'Execution Unit/Peak Dynamic': 5.81006,
'Execution Unit/Register Files/Area': 0.570804,
'Execution Unit/Register Files/Floating Point RF/Area': 0.208131,
'Execution Unit/Register Files/Floating Point RF/Gate Leakage': 0.000232788,
'Execution Unit/Register Files/Floating Point RF/Peak Dynamic': 0.075186,
'Execution Unit/Register Files/Floating Point RF/Runtime Dynamic': 0.00684429,
'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage': 0.00399698,
'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage with power gating': 0.00176968,
'Execution Unit/Register Files/Gate Leakage': 0.000622708,
'Execution Unit/Register Files/Integer RF/Area': 0.362673,
'Execution Unit/Register Files/Integer RF/Gate Leakage': 0.00038992,
'Execution Unit/Register Files/Integer RF/Peak Dynamic': 0.075131,
'Execution Unit/Register Files/Integer RF/Runtime Dynamic': 0.0506177,
'Execution Unit/Register Files/Integer RF/Subthreshold Leakage': 0.00614175,
'Execution Unit/Register Files/Integer RF/Subthreshold Leakage with power gating': 0.00246675,
'Execution Unit/Register Files/Peak Dynamic': 0.150317,
'Execution Unit/Register Files/Runtime Dynamic': 0.057462,
'Execution Unit/Register Files/Subthreshold Leakage': 0.0101387,
'Execution Unit/Register Files/Subthreshold Leakage with power gating': 0.00423643,
'Execution Unit/Results Broadcast Bus/Area Overhead': 0.0442632,
'Execution Unit/Results Broadcast Bus/Gate Leakage': 0.00607074,
'Execution Unit/Results Broadcast Bus/Peak Dynamic': 0.200837,
'Execution Unit/Results Broadcast Bus/Runtime Dynamic': 0.549203,
'Execution Unit/Results Broadcast Bus/Subthreshold Leakage': 0.0920413,
'Execution Unit/Results Broadcast Bus/Subthreshold Leakage with power gating': 0.0345155,
'Execution Unit/Runtime Dynamic': 1.97329,
'Execution Unit/Subthreshold Leakage': 1.83518,
'Execution Unit/Subthreshold Leakage with power gating': 0.709678,
'Gate Leakage': 0.372997,
'Instruction Fetch Unit/Area': 5.86007,
'Instruction Fetch Unit/Branch Predictor/Area': 0.138516,
'Instruction Fetch Unit/Branch Predictor/Chooser/Area': 0.0435221,
'Instruction Fetch Unit/Branch Predictor/Chooser/Gate Leakage': 0.000278362,
'Instruction Fetch Unit/Branch Predictor/Chooser/Peak Dynamic': 0.0168831,
'Instruction Fetch Unit/Branch Predictor/Chooser/Runtime Dynamic': 0.000128293,
'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage': 0.00759719,
'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage with power gating': 0.0039236,
'Instruction Fetch Unit/Branch Predictor/Gate Leakage': 0.000757657,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Area': 0.0435221,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Gate Leakage': 0.000278362,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Peak Dynamic': 0.0168831,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Runtime Dynamic': 0.000128293,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage': 0.00759719,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage with power gating': 0.0039236,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Area': 0.0257064,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Gate Leakage': 0.000154548,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Peak Dynamic': 0.0142575,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Runtime Dynamic': 0.000111033,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage': 0.00384344,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage with power gating': 0.00198631,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Area': 0.0151917,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Gate Leakage': 8.00196e-05,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Peak Dynamic': 0.00527447,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Runtime Dynamic': 4.25945e-05,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage': 0.00181347,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage with power gating': 0.000957045,
'Instruction Fetch Unit/Branch Predictor/Peak Dynamic': 0.0597838,
'Instruction Fetch Unit/Branch Predictor/RAS/Area': 0.0105732,
'Instruction Fetch Unit/Branch Predictor/RAS/Gate Leakage': 4.63858e-05,
'Instruction Fetch Unit/Branch Predictor/RAS/Peak Dynamic': 0.0117602,
'Instruction Fetch Unit/Branch Predictor/RAS/Runtime Dynamic': 0.000727128,
'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage': 0.000932505,
'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage with power gating': 0.000494733,
'Instruction Fetch Unit/Branch Predictor/Runtime Dynamic': 0.00109475,
'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage': 0.0199703,
'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage with power gating': 0.0103282,
'Instruction Fetch Unit/Branch Target Buffer/Area': 0.64954,
'Instruction Fetch Unit/Branch Target Buffer/Gate Leakage': 0.00272758,
'Instruction Fetch Unit/Branch Target Buffer/Peak Dynamic': 0.177867,
'Instruction Fetch Unit/Branch Target Buffer/Runtime Dynamic': 0.00125542,
'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage': 0.0811682,
'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage with power gating': 0.0435357,
'Instruction Fetch Unit/Gate Leakage': 0.0590479,
'Instruction Fetch Unit/Instruction Buffer/Area': 0.0226323,
'Instruction Fetch Unit/Instruction Buffer/Gate Leakage': 6.83558e-05,
'Instruction Fetch Unit/Instruction Buffer/Peak Dynamic': 0.606827,
'Instruction Fetch Unit/Instruction Buffer/Runtime Dynamic': 0.0486601,
'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage': 0.00151885,
'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage with power gating': 0.000701682,
'Instruction Fetch Unit/Instruction Cache/Area': 3.14635,
'Instruction Fetch Unit/Instruction Cache/Gate Leakage': 0.029931,
'Instruction Fetch Unit/Instruction Cache/Peak Dynamic': 3.0952,
'Instruction Fetch Unit/Instruction Cache/Runtime Dynamic': 0.117729,
'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage': 0.367022,
'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage with power gating': 0.180386,
'Instruction Fetch Unit/Instruction Decoder/Area': 1.85799,
'Instruction Fetch Unit/Instruction Decoder/Gate Leakage': 0.0222493,
'Instruction Fetch Unit/Instruction Decoder/Peak Dynamic': 1.37404,
'Instruction Fetch Unit/Instruction Decoder/Runtime Dynamic': 0.165272,
'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage': 0.442943,
'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage with power gating': 0.166104,
'Instruction Fetch Unit/Peak Dynamic': 5.46623,
'Instruction Fetch Unit/Runtime Dynamic': 0.334011,
'Instruction Fetch Unit/Subthreshold Leakage': 0.932587,
'Instruction Fetch Unit/Subthreshold Leakage with power gating': 0.408542,
'L2/Area': 4.53318,
'L2/Gate Leakage': 0.015464,
'L2/Peak Dynamic': 0.130932,
'L2/Runtime Dynamic': 0.037829,
'L2/Subthreshold Leakage': 0.834142,
'L2/Subthreshold Leakage with power gating': 0.401066,
'Load Store Unit/Area': 8.80969,
'Load Store Unit/Data Cache/Area': 6.84535,
'Load Store Unit/Data Cache/Gate Leakage': 0.0279261,
'Load Store Unit/Data Cache/Peak Dynamic': 3.29559,
'Load Store Unit/Data Cache/Runtime Dynamic': 1.04058,
'Load Store Unit/Data Cache/Subthreshold Leakage': 0.527675,
'Load Store Unit/Data Cache/Subthreshold Leakage with power gating': 0.25085,
'Load Store Unit/Gate Leakage': 0.0351387,
'Load Store Unit/LoadQ/Area': 0.0836782,
'Load Store Unit/LoadQ/Gate Leakage': 0.00059896,
'Load Store Unit/LoadQ/Peak Dynamic': 0.0665964,
'Load Store Unit/LoadQ/Runtime Dynamic': 0.0665963,
'Load Store Unit/LoadQ/Subthreshold Leakage': 0.00941961,
'Load Store Unit/LoadQ/Subthreshold Leakage with power gating': 0.00536918,
'Load Store Unit/Peak Dynamic': 3.61135,
'Load Store Unit/Runtime Dynamic': 1.4356,
'Load Store Unit/StoreQ/Area': 0.322079,
'Load Store Unit/StoreQ/Gate Leakage': 0.00329971,
'Load Store Unit/StoreQ/Peak Dynamic': 0.164215,
'Load Store Unit/StoreQ/Runtime Dynamic': 0.32843,
'Load Store Unit/StoreQ/Subthreshold Leakage': 0.0345621,
'Load Store Unit/StoreQ/Subthreshold Leakage with power gating': 0.0197004,
'Load Store Unit/Subthreshold Leakage': 0.591622,
'Load Store Unit/Subthreshold Leakage with power gating': 0.283406,
'Memory Management Unit/Area': 0.434579,
'Memory Management Unit/Dtlb/Area': 0.0879726,
'Memory Management Unit/Dtlb/Gate Leakage': 0.00088729,
'Memory Management Unit/Dtlb/Peak Dynamic': 0.0582806,
'Memory Management Unit/Dtlb/Runtime Dynamic': 0.0602397,
'Memory Management Unit/Dtlb/Subthreshold Leakage': 0.0155699,
'Memory Management Unit/Dtlb/Subthreshold Leakage with power gating': 0.00887485,
'Memory Management Unit/Gate Leakage': 0.00813591,
'Memory Management Unit/Itlb/Area': 0.301552,
'Memory Management Unit/Itlb/Gate Leakage': 0.00393464,
'Memory Management Unit/Itlb/Peak Dynamic': 0.192448,
'Memory Management Unit/Itlb/Runtime Dynamic': 0.0193218,
'Memory Management Unit/Itlb/Subthreshold Leakage': 0.0413758,
'Memory Management Unit/Itlb/Subthreshold Leakage with power gating': 0.0235842,
'Memory Management Unit/Peak Dynamic': 0.451703,
'Memory Management Unit/Runtime Dynamic': 0.0795615,
'Memory Management Unit/Subthreshold Leakage': 0.0769113,
'Memory Management Unit/Subthreshold Leakage with power gating': 0.0399462,
'Peak Dynamic': 20.032,
'Renaming Unit/Area': 0.369768,
'Renaming Unit/FP Front End RAT/Area': 0.168486,
'Renaming Unit/FP Front End RAT/Gate Leakage': 0.00489731,
'Renaming Unit/FP Front End RAT/Peak Dynamic': 3.33511,
'Renaming Unit/FP Front End RAT/Runtime Dynamic': 0.262307,
'Renaming Unit/FP Front End RAT/Subthreshold Leakage': 0.0437281,
'Renaming Unit/FP Front End RAT/Subthreshold Leakage with power gating': 0.024925,
'Renaming Unit/Free List/Area': 0.0414755,
'Renaming Unit/Free List/Gate Leakage': 4.15911e-05,
'Renaming Unit/Free List/Peak Dynamic': 0.0401324,
'Renaming Unit/Free List/Runtime Dynamic': 0.0128108,
'Renaming Unit/Free List/Subthreshold Leakage': 0.000670426,
'Renaming Unit/Free List/Subthreshold Leakage with power gating': 0.000377987,
'Renaming Unit/Gate Leakage': 0.00863632,
'Renaming Unit/Int Front End RAT/Area': 0.114751,
'Renaming Unit/Int Front End RAT/Gate Leakage': 0.00038343,
'Renaming Unit/Int Front End RAT/Peak Dynamic': 0.86945,
'Renaming Unit/Int Front End RAT/Runtime Dynamic': 0.09465,
'Renaming Unit/Int Front End RAT/Subthreshold Leakage': 0.00611897,
'Renaming Unit/Int Front End RAT/Subthreshold Leakage with power gating': 0.00348781,
'Renaming Unit/Peak Dynamic': 4.56169,
'Renaming Unit/Runtime Dynamic': 0.369768,
'Renaming Unit/Subthreshold Leakage': 0.070483,
'Renaming Unit/Subthreshold Leakage with power gating': 0.0362779,
'Runtime Dynamic': 4.23006,
'Subthreshold Leakage': 6.21877,
'Subthreshold Leakage with power gating': 2.58311},
{'Area': 32.0201,
'Execution Unit/Area': 7.68434,
'Execution Unit/Complex ALUs/Area': 0.235435,
'Execution Unit/Complex ALUs/Gate Leakage': 0.0132646,
'Execution Unit/Complex ALUs/Peak Dynamic': 0.023245,
'Execution Unit/Complex ALUs/Runtime Dynamic': 0.220946,
'Execution Unit/Complex ALUs/Subthreshold Leakage': 0.20111,
'Execution Unit/Complex ALUs/Subthreshold Leakage with power gating': 0.0754163,
'Execution Unit/Floating Point Units/Area': 4.6585,
'Execution Unit/Floating Point Units/Gate Leakage': 0.0656156,
'Execution Unit/Floating Point Units/Peak Dynamic': 0.131656,
'Execution Unit/Floating Point Units/Runtime Dynamic': 0.304033,
'Execution Unit/Floating Point Units/Subthreshold Leakage': 0.994829,
'Execution Unit/Floating Point Units/Subthreshold Leakage with power gating': 0.373061,
'Execution Unit/Gate Leakage': 0.120359,
'Execution Unit/Instruction Scheduler/Area': 1.66526,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Area': 0.275653,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Gate Leakage': 0.000977433,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Peak Dynamic': 1.04181,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Runtime Dynamic': 0.0536798,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage': 0.0143453,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage with power gating': 0.00810519,
'Execution Unit/Instruction Scheduler/Gate Leakage': 0.00568913,
'Execution Unit/Instruction Scheduler/Instruction Window/Area': 0.805223,
'Execution Unit/Instruction Scheduler/Instruction Window/Gate Leakage': 0.00414562,
'Execution Unit/Instruction Scheduler/Instruction Window/Peak Dynamic': 1.6763,
'Execution Unit/Instruction Scheduler/Instruction Window/Runtime Dynamic': 0.0865835,
'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage': 0.0625755,
'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage with power gating': 0.0355964,
'Execution Unit/Instruction Scheduler/Peak Dynamic': 3.82262,
'Execution Unit/Instruction Scheduler/ROB/Area': 0.584388,
'Execution Unit/Instruction Scheduler/ROB/Gate Leakage': 0.00056608,
'Execution Unit/Instruction Scheduler/ROB/Peak Dynamic': 1.10451,
'Execution Unit/Instruction Scheduler/ROB/Runtime Dynamic': 0.0437044,
'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage': 0.00906853,
'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage with power gating': 0.00364446,
'Execution Unit/Instruction Scheduler/Runtime Dynamic': 0.183968,
'Execution Unit/Instruction Scheduler/Subthreshold Leakage': 0.0859892,
'Execution Unit/Instruction Scheduler/Subthreshold Leakage with power gating': 0.047346,
'Execution Unit/Integer ALUs/Area': 0.47087,
'Execution Unit/Integer ALUs/Gate Leakage': 0.0265291,
'Execution Unit/Integer ALUs/Peak Dynamic': 0.0412103,
'Execution Unit/Integer ALUs/Runtime Dynamic': 0.101344,
'Execution Unit/Integer ALUs/Subthreshold Leakage': 0.40222,
'Execution Unit/Integer ALUs/Subthreshold Leakage with power gating': 0.150833,
'Execution Unit/Peak Dynamic': 4.13179,
'Execution Unit/Register Files/Area': 0.570804,
'Execution Unit/Register Files/Floating Point RF/Area': 0.208131,
'Execution Unit/Register Files/Floating Point RF/Gate Leakage': 0.000232788,
'Execution Unit/Register Files/Floating Point RF/Peak Dynamic': 0.0248725,
'Execution Unit/Register Files/Floating Point RF/Runtime Dynamic': 0.00225157,
'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage': 0.00399698,
'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage with power gating': 0.00176968,
'Execution Unit/Register Files/Gate Leakage': 0.000622708,
'Execution Unit/Register Files/Integer RF/Area': 0.362673,
'Execution Unit/Register Files/Integer RF/Gate Leakage': 0.00038992,
'Execution Unit/Register Files/Integer RF/Peak Dynamic': 0.0247354,
'Execution Unit/Register Files/Integer RF/Runtime Dynamic': 0.0166518,
'Execution Unit/Register Files/Integer RF/Subthreshold Leakage': 0.00614175,
'Execution Unit/Register Files/Integer RF/Subthreshold Leakage with power gating': 0.00246675,
'Execution Unit/Register Files/Peak Dynamic': 0.049608,
'Execution Unit/Register Files/Runtime Dynamic': 0.0189033,
'Execution Unit/Register Files/Subthreshold Leakage': 0.0101387,
'Execution Unit/Register Files/Subthreshold Leakage with power gating': 0.00423643,
'Execution Unit/Results Broadcast Bus/Area Overhead': 0.0390912,
'Execution Unit/Results Broadcast Bus/Gate Leakage': 0.00537402,
'Execution Unit/Results Broadcast Bus/Peak Dynamic': 0.0579232,
'Execution Unit/Results Broadcast Bus/Runtime Dynamic': 0.15908,
'Execution Unit/Results Broadcast Bus/Subthreshold Leakage': 0.081478,
'Execution Unit/Results Broadcast Bus/Subthreshold Leakage with power gating': 0.0305543,
'Execution Unit/Runtime Dynamic': 0.988274,
'Execution Unit/Subthreshold Leakage': 1.79543,
'Execution Unit/Subthreshold Leakage with power gating': 0.688821,
'Gate Leakage': 0.368936,
'Instruction Fetch Unit/Area': 5.85939,
'Instruction Fetch Unit/Branch Predictor/Area': 0.138516,
'Instruction Fetch Unit/Branch Predictor/Chooser/Area': 0.0435221,
'Instruction Fetch Unit/Branch Predictor/Chooser/Gate Leakage': 0.000278362,
'Instruction Fetch Unit/Branch Predictor/Chooser/Peak Dynamic': 0.0168831,
'Instruction Fetch Unit/Branch Predictor/Chooser/Runtime Dynamic': 4.39474e-05,
'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage': 0.00759719,
'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage with power gating': 0.0039236,
'Instruction Fetch Unit/Branch Predictor/Gate Leakage': 0.000757657,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Area': 0.0435221,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Gate Leakage': 0.000278362,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Peak Dynamic': 0.0168831,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Runtime Dynamic': 4.39474e-05,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage': 0.00759719,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage with power gating': 0.0039236,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Area': 0.0257064,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Gate Leakage': 0.000154548,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Peak Dynamic': 0.0142575,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Runtime Dynamic': 3.84241e-05,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage': 0.00384344,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage with power gating': 0.00198631,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Area': 0.0151917,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Gate Leakage': 8.00196e-05,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Peak Dynamic': 0.00527447,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Runtime Dynamic': 1.49544e-05,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage': 0.00181347,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage with power gating': 0.000957045,
'Instruction Fetch Unit/Branch Predictor/Peak Dynamic': 0.0597838,
'Instruction Fetch Unit/Branch Predictor/RAS/Area': 0.0105732,
'Instruction Fetch Unit/Branch Predictor/RAS/Gate Leakage': 4.63858e-05,
'Instruction Fetch Unit/Branch Predictor/RAS/Peak Dynamic': 0.0117602,
'Instruction Fetch Unit/Branch Predictor/RAS/Runtime Dynamic': 0.000239204,
'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage': 0.000932505,
'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage with power gating': 0.000494733,
'Instruction Fetch Unit/Branch Predictor/Runtime Dynamic': 0.000365523,
'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage': 0.0199703,
'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage with power gating': 0.0103282,
'Instruction Fetch Unit/Branch Target Buffer/Area': 0.64954,
'Instruction Fetch Unit/Branch Target Buffer/Gate Leakage': 0.00272758,
'Instruction Fetch Unit/Branch Target Buffer/Peak Dynamic': 0.177867,
'Instruction Fetch Unit/Branch Target Buffer/Runtime Dynamic': 0.000416149,
'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage': 0.0811682,
'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage with power gating': 0.0435357,
'Instruction Fetch Unit/Gate Leakage': 0.0589979,
'Instruction Fetch Unit/Instruction Buffer/Area': 0.0226323,
'Instruction Fetch Unit/Instruction Buffer/Gate Leakage': 6.83558e-05,
'Instruction Fetch Unit/Instruction Buffer/Peak Dynamic': 0.606827,
'Instruction Fetch Unit/Instruction Buffer/Runtime Dynamic': 0.0160078,
'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage': 0.00151885,
'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage with power gating': 0.000701682,
'Instruction Fetch Unit/Instruction Cache/Area': 3.14635,
'Instruction Fetch Unit/Instruction Cache/Gate Leakage': 0.029931,
'Instruction Fetch Unit/Instruction Cache/Peak Dynamic': 1.01823,
'Instruction Fetch Unit/Instruction Cache/Runtime Dynamic': 0.038844,
'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage': 0.367022,
'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage with power gating': 0.180386,
'Instruction Fetch Unit/Instruction Decoder/Area': 1.85799,
'Instruction Fetch Unit/Instruction Decoder/Gate Leakage': 0.0222493,
'Instruction Fetch Unit/Instruction Decoder/Peak Dynamic': 1.37404,
'Instruction Fetch Unit/Instruction Decoder/Runtime Dynamic': 0.0543695,
'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage': 0.442943,
'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage with power gating': 0.166104,
'Instruction Fetch Unit/Peak Dynamic': 3.28616,
'Instruction Fetch Unit/Runtime Dynamic': 0.110003,
'Instruction Fetch Unit/Subthreshold Leakage': 0.932286,
'Instruction Fetch Unit/Subthreshold Leakage with power gating': 0.40843,
'L2/Area': 4.53318,
'L2/Gate Leakage': 0.015464,
'L2/Peak Dynamic': 0.0427196,
'L2/Runtime Dynamic': 0.0124854,
'L2/Subthreshold Leakage': 0.834142,
'L2/Subthreshold Leakage with power gating': 0.401066,
'Load Store Unit/Area': 8.80901,
'Load Store Unit/Data Cache/Area': 6.84535,
'Load Store Unit/Data Cache/Gate Leakage': 0.0279261,
'Load Store Unit/Data Cache/Peak Dynamic': 1.9148,
'Load Store Unit/Data Cache/Runtime Dynamic': 0.342833,
'Load Store Unit/Data Cache/Subthreshold Leakage': 0.527675,
'Load Store Unit/Data Cache/Subthreshold Leakage with power gating': 0.25085,
'Load Store Unit/Gate Leakage': 0.0350888,
'Load Store Unit/LoadQ/Area': 0.0836782,
'Load Store Unit/LoadQ/Gate Leakage': 0.00059896,
'Load Store Unit/LoadQ/Peak Dynamic': 0.0219246,
'Load Store Unit/LoadQ/Runtime Dynamic': 0.0219246,
'Load Store Unit/LoadQ/Subthreshold Leakage': 0.00941961,
'Load Store Unit/LoadQ/Subthreshold Leakage with power gating': 0.00536918,
'Load Store Unit/Peak Dynamic': 2.01833,
'Load Store Unit/Runtime Dynamic': 0.472882,
'Load Store Unit/StoreQ/Area': 0.322079,
'Load Store Unit/StoreQ/Gate Leakage': 0.00329971,
'Load Store Unit/StoreQ/Peak Dynamic': 0.0540624,
'Load Store Unit/StoreQ/Runtime Dynamic': 0.108125,
'Load Store Unit/StoreQ/Subthreshold Leakage': 0.0345621,
'Load Store Unit/StoreQ/Subthreshold Leakage with power gating': 0.0197004,
'Load Store Unit/Subthreshold Leakage': 0.591321,
'Load Store Unit/Subthreshold Leakage with power gating': 0.283293,
'Memory Management Unit/Area': 0.4339,
'Memory Management Unit/Dtlb/Area': 0.0879726,
'Memory Management Unit/Dtlb/Gate Leakage': 0.00088729,
'Memory Management Unit/Dtlb/Peak Dynamic': 0.0191869,
'Memory Management Unit/Dtlb/Runtime Dynamic': 0.0198261,
'Memory Management Unit/Dtlb/Subthreshold Leakage': 0.0155699,
'Memory Management Unit/Dtlb/Subthreshold Leakage with power gating': 0.00887485,
'Memory Management Unit/Gate Leakage': 0.00808595,
'Memory Management Unit/Itlb/Area': 0.301552,
'Memory Management Unit/Itlb/Gate Leakage': 0.00393464,
'Memory Management Unit/Itlb/Peak Dynamic': 0.06331,
'Memory Management Unit/Itlb/Runtime Dynamic': 0.0063751,
'Memory Management Unit/Itlb/Subthreshold Leakage': 0.0413758,
'Memory Management Unit/Itlb/Subthreshold Leakage with power gating': 0.0235842,
'Memory Management Unit/Peak Dynamic': 0.252379,
'Memory Management Unit/Runtime Dynamic': 0.0262012,
'Memory Management Unit/Subthreshold Leakage': 0.0766103,
'Memory Management Unit/Subthreshold Leakage with power gating': 0.0398333,
'Peak Dynamic': 13.3209,
'Renaming Unit/Area': 0.303608,
'Renaming Unit/FP Front End RAT/Area': 0.131045,
'Renaming Unit/FP Front End RAT/Gate Leakage': 0.00351123,
'Renaming Unit/FP Front End RAT/Peak Dynamic': 2.51468,
'Renaming Unit/FP Front End RAT/Runtime Dynamic': 0.0654283,
'Renaming Unit/FP Front End RAT/Subthreshold Leakage': 0.0308571,
'Renaming Unit/FP Front End RAT/Subthreshold Leakage with power gating': 0.0175885,
'Renaming Unit/Free List/Area': 0.0340654,
'Renaming Unit/Free List/Gate Leakage': 2.5481e-05,
'Renaming Unit/Free List/Peak Dynamic': 0.0306032,
'Renaming Unit/Free List/Runtime Dynamic': 0.00321813,
'Renaming Unit/Free List/Subthreshold Leakage': 0.000370144,
'Renaming Unit/Free List/Subthreshold Leakage with power gating': 0.000201064,
'Renaming Unit/Gate Leakage': 0.00708398,
'Renaming Unit/Int Front End RAT/Area': 0.0941223,
'Renaming Unit/Int Front End RAT/Gate Leakage': 0.000283242,
'Renaming Unit/Int Front End RAT/Peak Dynamic': 0.731965,
'Renaming Unit/Int Front End RAT/Runtime Dynamic': 0.0264341,
'Renaming Unit/Int Front End RAT/Subthreshold Leakage': 0.00435488,
'Renaming Unit/Int Front End RAT/Subthreshold Leakage with power gating': 0.00248228,
'Renaming Unit/Peak Dynamic': 3.58947,
'Renaming Unit/Runtime Dynamic': 0.0950806,
'Renaming Unit/Subthreshold Leakage': 0.0552466,
'Renaming Unit/Subthreshold Leakage with power gating': 0.0276461,
'Runtime Dynamic': 1.70493,
'Subthreshold Leakage': 6.16288,
'Subthreshold Leakage with power gating': 2.55328},
{'Area': 32.0201,
'Execution Unit/Area': 7.68434,
'Execution Unit/Complex ALUs/Area': 0.235435,
'Execution Unit/Complex ALUs/Gate Leakage': 0.0132646,
'Execution Unit/Complex ALUs/Peak Dynamic': 0.0232413,
'Execution Unit/Complex ALUs/Runtime Dynamic': 0.220943,
'Execution Unit/Complex ALUs/Subthreshold Leakage': 0.20111,
'Execution Unit/Complex ALUs/Subthreshold Leakage with power gating': 0.0754163,
'Execution Unit/Floating Point Units/Area': 4.6585,
'Execution Unit/Floating Point Units/Gate Leakage': 0.0656156,
'Execution Unit/Floating Point Units/Peak Dynamic': 0.13164,
'Execution Unit/Floating Point Units/Runtime Dynamic': 0.304033,
'Execution Unit/Floating Point Units/Subthreshold Leakage': 0.994829,
'Execution Unit/Floating Point Units/Subthreshold Leakage with power gating': 0.373061,
'Execution Unit/Gate Leakage': 0.120359,
'Execution Unit/Instruction Scheduler/Area': 1.66526,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Area': 0.275653,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Gate Leakage': 0.000977433,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Peak Dynamic': 1.04181,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Runtime Dynamic': 0.0536686,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage': 0.0143453,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage with power gating': 0.00810519,
'Execution Unit/Instruction Scheduler/Gate Leakage': 0.00568913,
'Execution Unit/Instruction Scheduler/Instruction Window/Area': 0.805223,
'Execution Unit/Instruction Scheduler/Instruction Window/Gate Leakage': 0.00414562,
'Execution Unit/Instruction Scheduler/Instruction Window/Peak Dynamic': 1.6763,
'Execution Unit/Instruction Scheduler/Instruction Window/Runtime Dynamic': 0.0865655,
'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage': 0.0625755,
'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage with power gating': 0.0355964,
'Execution Unit/Instruction Scheduler/Peak Dynamic': 3.82262,
'Execution Unit/Instruction Scheduler/ROB/Area': 0.584388,
'Execution Unit/Instruction Scheduler/ROB/Gate Leakage': 0.00056608,
'Execution Unit/Instruction Scheduler/ROB/Peak Dynamic': 1.10451,
'Execution Unit/Instruction Scheduler/ROB/Runtime Dynamic': 0.0436953,
'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage': 0.00906853,
'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage with power gating': 0.00364446,
'Execution Unit/Instruction Scheduler/Runtime Dynamic': 0.183929,
'Execution Unit/Instruction Scheduler/Subthreshold Leakage': 0.0859892,
'Execution Unit/Instruction Scheduler/Subthreshold Leakage with power gating': 0.047346,
'Execution Unit/Integer ALUs/Area': 0.47087,
'Execution Unit/Integer ALUs/Gate Leakage': 0.0265291,
'Execution Unit/Integer ALUs/Peak Dynamic': 0.0411987,
'Execution Unit/Integer ALUs/Runtime Dynamic': 0.101344,
'Execution Unit/Integer ALUs/Subthreshold Leakage': 0.40222,
'Execution Unit/Integer ALUs/Subthreshold Leakage with power gating': 0.150833,
'Execution Unit/Peak Dynamic': 4.13174,
'Execution Unit/Register Files/Area': 0.570804,
'Execution Unit/Register Files/Floating Point RF/Area': 0.208131,
'Execution Unit/Register Files/Floating Point RF/Gate Leakage': 0.000232788,
'Execution Unit/Register Files/Floating Point RF/Peak Dynamic': 0.0248697,
'Execution Unit/Register Files/Floating Point RF/Runtime Dynamic': 0.0022511,
'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage': 0.00399698,
'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage with power gating': 0.00176968,
'Execution Unit/Register Files/Gate Leakage': 0.000622708,
'Execution Unit/Register Files/Integer RF/Area': 0.362673,
'Execution Unit/Register Files/Integer RF/Gate Leakage': 0.00038992,
'Execution Unit/Register Files/Integer RF/Peak Dynamic': 0.0247301,
'Execution Unit/Register Files/Integer RF/Runtime Dynamic': 0.0166483,
'Execution Unit/Register Files/Integer RF/Subthreshold Leakage': 0.00614175,
'Execution Unit/Register Files/Integer RF/Subthreshold Leakage with power gating': 0.00246675,
'Execution Unit/Register Files/Peak Dynamic': 0.0495998,
'Execution Unit/Register Files/Runtime Dynamic': 0.0188994,
'Execution Unit/Register Files/Subthreshold Leakage': 0.0101387,
'Execution Unit/Register Files/Subthreshold Leakage with power gating': 0.00423643,
'Execution Unit/Results Broadcast Bus/Area Overhead': 0.0390912,
'Execution Unit/Results Broadcast Bus/Gate Leakage': 0.00537402,
'Execution Unit/Results Broadcast Bus/Peak Dynamic': 0.057911,
'Execution Unit/Results Broadcast Bus/Runtime Dynamic': 0.15905,
'Execution Unit/Results Broadcast Bus/Subthreshold Leakage': 0.081478,
'Execution Unit/Results Broadcast Bus/Subthreshold Leakage with power gating': 0.0305543,
'Execution Unit/Runtime Dynamic': 0.9882,
'Execution Unit/Subthreshold Leakage': 1.79543,
'Execution Unit/Subthreshold Leakage with power gating': 0.688821,
'Gate Leakage': 0.368936,
'Instruction Fetch Unit/Area': 5.85939,
'Instruction Fetch Unit/Branch Predictor/Area': 0.138516,
'Instruction Fetch Unit/Branch Predictor/Chooser/Area': 0.0435221,
'Instruction Fetch Unit/Branch Predictor/Chooser/Gate Leakage': 0.000278362,
'Instruction Fetch Unit/Branch Predictor/Chooser/Peak Dynamic': 0.0168831,
'Instruction Fetch Unit/Branch Predictor/Chooser/Runtime Dynamic': 4.39436e-05,
'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage': 0.00759719,
'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage with power gating': 0.0039236,
'Instruction Fetch Unit/Branch Predictor/Gate Leakage': 0.000757657,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Area': 0.0435221,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Gate Leakage': 0.000278362,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Peak Dynamic': 0.0168831,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Runtime Dynamic': 4.39436e-05,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage': 0.00759719,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage with power gating': 0.0039236,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Area': 0.0257064,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Gate Leakage': 0.000154548,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Peak Dynamic': 0.0142575,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Runtime Dynamic': 3.84208e-05,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage': 0.00384344,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage with power gating': 0.00198631,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Area': 0.0151917,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Gate Leakage': 8.00196e-05,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Peak Dynamic': 0.00527447,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Runtime Dynamic': 1.49532e-05,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage': 0.00181347,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage with power gating': 0.000957045,
'Instruction Fetch Unit/Branch Predictor/Peak Dynamic': 0.0597838,
'Instruction Fetch Unit/Branch Predictor/RAS/Area': 0.0105732,
'Instruction Fetch Unit/Branch Predictor/RAS/Gate Leakage': 4.63858e-05,
'Instruction Fetch Unit/Branch Predictor/RAS/Peak Dynamic': 0.0117602,
'Instruction Fetch Unit/Branch Predictor/RAS/Runtime Dynamic': 0.000239154,
'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage': 0.000932505,
'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage with power gating': 0.000494733,
'Instruction Fetch Unit/Branch Predictor/Runtime Dynamic': 0.000365462,
'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage': 0.0199703,
'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage with power gating': 0.0103282,
'Instruction Fetch Unit/Branch Target Buffer/Area': 0.64954,
'Instruction Fetch Unit/Branch Target Buffer/Gate Leakage': 0.00272758,
'Instruction Fetch Unit/Branch Target Buffer/Peak Dynamic': 0.177867,
'Instruction Fetch Unit/Branch Target Buffer/Runtime Dynamic': 0.000416112,
'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage': 0.0811682,
'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage with power gating': 0.0435357,
'Instruction Fetch Unit/Gate Leakage': 0.0589979,
'Instruction Fetch Unit/Instruction Buffer/Area': 0.0226323,
'Instruction Fetch Unit/Instruction Buffer/Gate Leakage': 6.83558e-05,
'Instruction Fetch Unit/Instruction Buffer/Peak Dynamic': 0.606827,
'Instruction Fetch Unit/Instruction Buffer/Runtime Dynamic': 0.0160044,
'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage': 0.00151885,
'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage with power gating': 0.000701682,
'Instruction Fetch Unit/Instruction Cache/Area': 3.14635,
'Instruction Fetch Unit/Instruction Cache/Gate Leakage': 0.029931,
'Instruction Fetch Unit/Instruction Cache/Peak Dynamic': 1.01802,
'Instruction Fetch Unit/Instruction Cache/Runtime Dynamic': 0.0388374,
'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage': 0.367022,
'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage with power gating': 0.180386,
'Instruction Fetch Unit/Instruction Decoder/Area': 1.85799,
'Instruction Fetch Unit/Instruction Decoder/Gate Leakage': 0.0222493,
'Instruction Fetch Unit/Instruction Decoder/Peak Dynamic': 1.37404,
'Instruction Fetch Unit/Instruction Decoder/Runtime Dynamic': 0.0543582,
'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage': 0.442943,
'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage with power gating': 0.166104,
'Instruction Fetch Unit/Peak Dynamic': 3.28594,
'Instruction Fetch Unit/Runtime Dynamic': 0.109982,
'Instruction Fetch Unit/Subthreshold Leakage': 0.932286,
'Instruction Fetch Unit/Subthreshold Leakage with power gating': 0.40843,
'L2/Area': 4.53318,
'L2/Gate Leakage': 0.015464,
'L2/Peak Dynamic': 0.0427043,
'L2/Runtime Dynamic': 0.0124827,
'L2/Subthreshold Leakage': 0.834142,
'L2/Subthreshold Leakage with power gating': 0.401066,
'Load Store Unit/Area': 8.80901,
'Load Store Unit/Data Cache/Area': 6.84535,
'Load Store Unit/Data Cache/Gate Leakage': 0.0279261,
'Load Store Unit/Data Cache/Peak Dynamic': 1.91467,
'Load Store Unit/Data Cache/Runtime Dynamic': 0.342768,
'Load Store Unit/Data Cache/Subthreshold Leakage': 0.527675,
'Load Store Unit/Data Cache/Subthreshold Leakage with power gating': 0.25085,
'Load Store Unit/Gate Leakage': 0.0350888,
'Load Store Unit/LoadQ/Area': 0.0836782,
'Load Store Unit/LoadQ/Gate Leakage': 0.00059896,
'Load Store Unit/LoadQ/Peak Dynamic': 0.0219204,
'Load Store Unit/LoadQ/Runtime Dynamic': 0.0219203,
'Load Store Unit/LoadQ/Subthreshold Leakage': 0.00941961,
'Load Store Unit/LoadQ/Subthreshold Leakage with power gating': 0.00536918,
'Load Store Unit/Peak Dynamic': 2.01818,
'Load Store Unit/Runtime Dynamic': 0.472792,
'Load Store Unit/StoreQ/Area': 0.322079,
'Load Store Unit/StoreQ/Gate Leakage': 0.00329971,
'Load Store Unit/StoreQ/Peak Dynamic': 0.054052,
'Load Store Unit/StoreQ/Runtime Dynamic': 0.108104,
'Load Store Unit/StoreQ/Subthreshold Leakage': 0.0345621,
'Load Store Unit/StoreQ/Subthreshold Leakage with power gating': 0.0197004,
'Load Store Unit/Subthreshold Leakage': 0.591321,
'Load Store Unit/Subthreshold Leakage with power gating': 0.283293,
'Memory Management Unit/Area': 0.4339,
'Memory Management Unit/Dtlb/Area': 0.0879726,
'Memory Management Unit/Dtlb/Gate Leakage': 0.00088729,
'Memory Management Unit/Dtlb/Peak Dynamic': 0.0191832,
'Memory Management Unit/Dtlb/Runtime Dynamic': 0.0198221,
'Memory Management Unit/Dtlb/Subthreshold Leakage': 0.0155699,
'Memory Management Unit/Dtlb/Subthreshold Leakage with power gating': 0.00887485,
'Memory Management Unit/Gate Leakage': 0.00808595,
'Memory Management Unit/Itlb/Area': 0.301552,
'Memory Management Unit/Itlb/Gate Leakage': 0.00393464,
'Memory Management Unit/Itlb/Peak Dynamic': 0.0632968,
'Memory Management Unit/Itlb/Runtime Dynamic': 0.00637403,
'Memory Management Unit/Itlb/Subthreshold Leakage': 0.0413758,
'Memory Management Unit/Itlb/Subthreshold Leakage with power gating': 0.0235842,
'Memory Management Unit/Peak Dynamic': 0.252359,
'Memory Management Unit/Runtime Dynamic': 0.0261961,
'Memory Management Unit/Subthreshold Leakage': 0.0766103,
'Memory Management Unit/Subthreshold Leakage with power gating': 0.0398333,
'Peak Dynamic': 13.3204,
'Renaming Unit/Area': 0.303608,
'Renaming Unit/FP Front End RAT/Area': 0.131045,
'Renaming Unit/FP Front End RAT/Gate Leakage': 0.00351123,
'Renaming Unit/FP Front End RAT/Peak Dynamic': 2.51468,
'Renaming Unit/FP Front End RAT/Runtime Dynamic': 0.0654206,
'Renaming Unit/FP Front End RAT/Subthreshold Leakage': 0.0308571,
'Renaming Unit/FP Front End RAT/Subthreshold Leakage with power gating': 0.0175885,
'Renaming Unit/Free List/Area': 0.0340654,
'Renaming Unit/Free List/Gate Leakage': 2.5481e-05,
'Renaming Unit/Free List/Peak Dynamic': 0.0306032,
'Renaming Unit/Free List/Runtime Dynamic': 0.00321753,
'Renaming Unit/Free List/Subthreshold Leakage': 0.000370144,
'Renaming Unit/Free List/Subthreshold Leakage with power gating': 0.000201064,
'Renaming Unit/Gate Leakage': 0.00708398,
'Renaming Unit/Int Front End RAT/Area': 0.0941223,
'Renaming Unit/Int Front End RAT/Gate Leakage': 0.000283242,
'Renaming Unit/Int Front End RAT/Peak Dynamic': 0.731965,
'Renaming Unit/Int Front End RAT/Runtime Dynamic': 0.0264284,
'Renaming Unit/Int Front End RAT/Subthreshold Leakage': 0.00435488,
'Renaming Unit/Int Front End RAT/Subthreshold Leakage with power gating': 0.00248228,
'Renaming Unit/Peak Dynamic': 3.58947,
'Renaming Unit/Runtime Dynamic': 0.0950665,
'Renaming Unit/Subthreshold Leakage': 0.0552466,
'Renaming Unit/Subthreshold Leakage with power gating': 0.0276461,
'Runtime Dynamic': 1.70472,
'Subthreshold Leakage': 6.16288,
'Subthreshold Leakage with power gating': 2.55328},
{'Area': 32.0201,
'Execution Unit/Area': 7.68434,
'Execution Unit/Complex ALUs/Area': 0.235435,
'Execution Unit/Complex ALUs/Gate Leakage': 0.0132646,
'Execution Unit/Complex ALUs/Peak Dynamic': 0.0232413,
'Execution Unit/Complex ALUs/Runtime Dynamic': 0.220943,
'Execution Unit/Complex ALUs/Subthreshold Leakage': 0.20111,
'Execution Unit/Complex ALUs/Subthreshold Leakage with power gating': 0.0754163,
'Execution Unit/Floating Point Units/Area': 4.6585,
'Execution Unit/Floating Point Units/Gate Leakage': 0.0656156,
'Execution Unit/Floating Point Units/Peak Dynamic': 0.13164,
'Execution Unit/Floating Point Units/Runtime Dynamic': 0.304033,
'Execution Unit/Floating Point Units/Subthreshold Leakage': 0.994829,
'Execution Unit/Floating Point Units/Subthreshold Leakage with power gating': 0.373061,
'Execution Unit/Gate Leakage': 0.120359,
'Execution Unit/Instruction Scheduler/Area': 1.66526,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Area': 0.275653,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Gate Leakage': 0.000977433,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Peak Dynamic': 1.04181,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Runtime Dynamic': 0.0536692,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage': 0.0143453,
'Execution Unit/Instruction Scheduler/FP Instruction Window/Subthreshold Leakage with power gating': 0.00810519,
'Execution Unit/Instruction Scheduler/Gate Leakage': 0.00568913,
'Execution Unit/Instruction Scheduler/Instruction Window/Area': 0.805223,
'Execution Unit/Instruction Scheduler/Instruction Window/Gate Leakage': 0.00414562,
'Execution Unit/Instruction Scheduler/Instruction Window/Peak Dynamic': 1.6763,
'Execution Unit/Instruction Scheduler/Instruction Window/Runtime Dynamic': 0.0865663,
'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage': 0.0625755,
'Execution Unit/Instruction Scheduler/Instruction Window/Subthreshold Leakage with power gating': 0.0355964,
'Execution Unit/Instruction Scheduler/Peak Dynamic': 3.82262,
'Execution Unit/Instruction Scheduler/ROB/Area': 0.584388,
'Execution Unit/Instruction Scheduler/ROB/Gate Leakage': 0.00056608,
'Execution Unit/Instruction Scheduler/ROB/Peak Dynamic': 1.10451,
'Execution Unit/Instruction Scheduler/ROB/Runtime Dynamic': 0.0436958,
'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage': 0.00906853,
'Execution Unit/Instruction Scheduler/ROB/Subthreshold Leakage with power gating': 0.00364446,
'Execution Unit/Instruction Scheduler/Runtime Dynamic': 0.183931,
'Execution Unit/Instruction Scheduler/Subthreshold Leakage': 0.0859892,
'Execution Unit/Instruction Scheduler/Subthreshold Leakage with power gating': 0.047346,
'Execution Unit/Integer ALUs/Area': 0.47087,
'Execution Unit/Integer ALUs/Gate Leakage': 0.0265291,
'Execution Unit/Integer ALUs/Peak Dynamic': 0.0411987,
'Execution Unit/Integer ALUs/Runtime Dynamic': 0.101344,
'Execution Unit/Integer ALUs/Subthreshold Leakage': 0.40222,
'Execution Unit/Integer ALUs/Subthreshold Leakage with power gating': 0.150833,
'Execution Unit/Peak Dynamic': 4.13174,
'Execution Unit/Register Files/Area': 0.570804,
'Execution Unit/Register Files/Floating Point RF/Area': 0.208131,
'Execution Unit/Register Files/Floating Point RF/Gate Leakage': 0.000232788,
'Execution Unit/Register Files/Floating Point RF/Peak Dynamic': 0.0248697,
'Execution Unit/Register Files/Floating Point RF/Runtime Dynamic': 0.00225113,
'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage': 0.00399698,
'Execution Unit/Register Files/Floating Point RF/Subthreshold Leakage with power gating': 0.00176968,
'Execution Unit/Register Files/Gate Leakage': 0.000622708,
'Execution Unit/Register Files/Integer RF/Area': 0.362673,
'Execution Unit/Register Files/Integer RF/Gate Leakage': 0.00038992,
'Execution Unit/Register Files/Integer RF/Peak Dynamic': 0.0247301,
'Execution Unit/Register Files/Integer RF/Runtime Dynamic': 0.0166484,
'Execution Unit/Register Files/Integer RF/Subthreshold Leakage': 0.00614175,
'Execution Unit/Register Files/Integer RF/Subthreshold Leakage with power gating': 0.00246675,
'Execution Unit/Register Files/Peak Dynamic': 0.0495998,
'Execution Unit/Register Files/Runtime Dynamic': 0.0188996,
'Execution Unit/Register Files/Subthreshold Leakage': 0.0101387,
'Execution Unit/Register Files/Subthreshold Leakage with power gating': 0.00423643,
'Execution Unit/Results Broadcast Bus/Area Overhead': 0.0390912,
'Execution Unit/Results Broadcast Bus/Gate Leakage': 0.00537402,
'Execution Unit/Results Broadcast Bus/Peak Dynamic': 0.057911,
'Execution Unit/Results Broadcast Bus/Runtime Dynamic': 0.159052,
'Execution Unit/Results Broadcast Bus/Subthreshold Leakage': 0.081478,
'Execution Unit/Results Broadcast Bus/Subthreshold Leakage with power gating': 0.0305543,
'Execution Unit/Runtime Dynamic': 0.988203,
'Execution Unit/Subthreshold Leakage': 1.79543,
'Execution Unit/Subthreshold Leakage with power gating': 0.688821,
'Gate Leakage': 0.368936,
'Instruction Fetch Unit/Area': 5.85939,
'Instruction Fetch Unit/Branch Predictor/Area': 0.138516,
'Instruction Fetch Unit/Branch Predictor/Chooser/Area': 0.0435221,
'Instruction Fetch Unit/Branch Predictor/Chooser/Gate Leakage': 0.000278362,
'Instruction Fetch Unit/Branch Predictor/Chooser/Peak Dynamic': 0.0168831,
'Instruction Fetch Unit/Branch Predictor/Chooser/Runtime Dynamic': 4.39438e-05,
'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage': 0.00759719,
'Instruction Fetch Unit/Branch Predictor/Chooser/Subthreshold Leakage with power gating': 0.0039236,
'Instruction Fetch Unit/Branch Predictor/Gate Leakage': 0.000757657,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Area': 0.0435221,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Gate Leakage': 0.000278362,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Peak Dynamic': 0.0168831,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Runtime Dynamic': 4.39438e-05,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage': 0.00759719,
'Instruction Fetch Unit/Branch Predictor/Global Predictor/Subthreshold Leakage with power gating': 0.0039236,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Area': 0.0257064,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Gate Leakage': 0.000154548,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Peak Dynamic': 0.0142575,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Runtime Dynamic': 3.8421e-05,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage': 0.00384344,
'Instruction Fetch Unit/Branch Predictor/L1_Local Predictor/Subthreshold Leakage with power gating': 0.00198631,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Area': 0.0151917,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Gate Leakage': 8.00196e-05,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Peak Dynamic': 0.00527447,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Runtime Dynamic': 1.49532e-05,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage': 0.00181347,
'Instruction Fetch Unit/Branch Predictor/L2_Local Predictor/Subthreshold Leakage with power gating': 0.000957045,
'Instruction Fetch Unit/Branch Predictor/Peak Dynamic': 0.0597838,
'Instruction Fetch Unit/Branch Predictor/RAS/Area': 0.0105732,
'Instruction Fetch Unit/Branch Predictor/RAS/Gate Leakage': 4.63858e-05,
'Instruction Fetch Unit/Branch Predictor/RAS/Peak Dynamic': 0.0117602,
'Instruction Fetch Unit/Branch Predictor/RAS/Runtime Dynamic': 0.000239156,
'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage': 0.000932505,
'Instruction Fetch Unit/Branch Predictor/RAS/Subthreshold Leakage with power gating': 0.000494733,
'Instruction Fetch Unit/Branch Predictor/Runtime Dynamic': 0.000365465,
'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage': 0.0199703,
'Instruction Fetch Unit/Branch Predictor/Subthreshold Leakage with power gating': 0.0103282,
'Instruction Fetch Unit/Branch Target Buffer/Area': 0.64954,
'Instruction Fetch Unit/Branch Target Buffer/Gate Leakage': 0.00272758,
'Instruction Fetch Unit/Branch Target Buffer/Peak Dynamic': 0.177867,
'Instruction Fetch Unit/Branch Target Buffer/Runtime Dynamic': 0.000416114,
'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage': 0.0811682,
'Instruction Fetch Unit/Branch Target Buffer/Subthreshold Leakage with power gating': 0.0435357,
'Instruction Fetch Unit/Gate Leakage': 0.0589979,
'Instruction Fetch Unit/Instruction Buffer/Area': 0.0226323,
'Instruction Fetch Unit/Instruction Buffer/Gate Leakage': 6.83558e-05,
'Instruction Fetch Unit/Instruction Buffer/Peak Dynamic': 0.606827,
'Instruction Fetch Unit/Instruction Buffer/Runtime Dynamic': 0.0160046,
'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage': 0.00151885,
'Instruction Fetch Unit/Instruction Buffer/Subthreshold Leakage with power gating': 0.000701682,
'Instruction Fetch Unit/Instruction Cache/Area': 3.14635,
'Instruction Fetch Unit/Instruction Cache/Gate Leakage': 0.029931,
'Instruction Fetch Unit/Instruction Cache/Peak Dynamic': 1.01803,
'Instruction Fetch Unit/Instruction Cache/Runtime Dynamic': 0.0388377,
'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage': 0.367022,
'Instruction Fetch Unit/Instruction Cache/Subthreshold Leakage with power gating': 0.180386,
'Instruction Fetch Unit/Instruction Decoder/Area': 1.85799,
'Instruction Fetch Unit/Instruction Decoder/Gate Leakage': 0.0222493,
'Instruction Fetch Unit/Instruction Decoder/Peak Dynamic': 1.37404,
'Instruction Fetch Unit/Instruction Decoder/Runtime Dynamic': 0.0543588,
'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage': 0.442943,
'Instruction Fetch Unit/Instruction Decoder/Subthreshold Leakage with power gating': 0.166104,
'Instruction Fetch Unit/Peak Dynamic': 3.28595,
'Instruction Fetch Unit/Runtime Dynamic': 0.109983,
'Instruction Fetch Unit/Subthreshold Leakage': 0.932286,
'Instruction Fetch Unit/Subthreshold Leakage with power gating': 0.40843,
'L2/Area': 4.53318,
'L2/Gate Leakage': 0.015464,
'L2/Peak Dynamic': 0.0427043,
'L2/Runtime Dynamic': 0.0124828,
'L2/Subthreshold Leakage': 0.834142,
'L2/Subthreshold Leakage with power gating': 0.401066,
'Load Store Unit/Area': 8.80901,
'Load Store Unit/Data Cache/Area': 6.84535,
'Load Store Unit/Data Cache/Gate Leakage': 0.0279261,
'Load Store Unit/Data Cache/Peak Dynamic': 1.91468,
'Load Store Unit/Data Cache/Runtime Dynamic': 0.342771,
'Load Store Unit/Data Cache/Subthreshold Leakage': 0.527675,
'Load Store Unit/Data Cache/Subthreshold Leakage with power gating': 0.25085,
'Load Store Unit/Gate Leakage': 0.0350888,
'Load Store Unit/LoadQ/Area': 0.0836782,
'Load Store Unit/LoadQ/Gate Leakage': 0.00059896,
'Load Store Unit/LoadQ/Peak Dynamic': 0.0219206,
'Load Store Unit/LoadQ/Runtime Dynamic': 0.0219205,
'Load Store Unit/LoadQ/Subthreshold Leakage': 0.00941961,
'Load Store Unit/LoadQ/Subthreshold Leakage with power gating': 0.00536918,
'Load Store Unit/Peak Dynamic': 2.01819,
'Load Store Unit/Runtime Dynamic': 0.472796,
'Load Store Unit/StoreQ/Area': 0.322079,
'Load Store Unit/StoreQ/Gate Leakage': 0.00329971,
'Load Store Unit/StoreQ/Peak Dynamic': 0.0540526,
'Load Store Unit/StoreQ/Runtime Dynamic': 0.108105,
'Load Store Unit/StoreQ/Subthreshold Leakage': 0.0345621,
'Load Store Unit/StoreQ/Subthreshold Leakage with power gating': 0.0197004,
'Load Store Unit/Subthreshold Leakage': 0.591321,
'Load Store Unit/Subthreshold Leakage with power gating': 0.283293,
'Memory Management Unit/Area': 0.4339,
'Memory Management Unit/Dtlb/Area': 0.0879726,
'Memory Management Unit/Dtlb/Gate Leakage': 0.00088729,
'Memory Management Unit/Dtlb/Peak Dynamic': 0.0191834,
'Memory Management Unit/Dtlb/Runtime Dynamic': 0.0198223,
'Memory Management Unit/Dtlb/Subthreshold Leakage': 0.0155699,
'Memory Management Unit/Dtlb/Subthreshold Leakage with power gating': 0.00887485,
'Memory Management Unit/Gate Leakage': 0.00808595,
'Memory Management Unit/Itlb/Area': 0.301552,
'Memory Management Unit/Itlb/Gate Leakage': 0.00393464,
'Memory Management Unit/Itlb/Peak Dynamic': 0.0632972,
'Memory Management Unit/Itlb/Runtime Dynamic': 0.00637408,
'Memory Management Unit/Itlb/Subthreshold Leakage': 0.0413758,
'Memory Management Unit/Itlb/Subthreshold Leakage with power gating': 0.0235842,
'Memory Management Unit/Peak Dynamic': 0.25236,
'Memory Management Unit/Runtime Dynamic': 0.0261964,
'Memory Management Unit/Subthreshold Leakage': 0.0766103,
'Memory Management Unit/Subthreshold Leakage with power gating': 0.0398333,
'Peak Dynamic': 13.3204,
'Renaming Unit/Area': 0.303608,
'Renaming Unit/FP Front End RAT/Area': 0.131045,
'Renaming Unit/FP Front End RAT/Gate Leakage': 0.00351123,
'Renaming Unit/FP Front End RAT/Peak Dynamic': 2.51468,
'Renaming Unit/FP Front End RAT/Runtime Dynamic': 0.0654209,
'Renaming Unit/FP Front End RAT/Subthreshold Leakage': 0.0308571,
'Renaming Unit/FP Front End RAT/Subthreshold Leakage with power gating': 0.0175885,
'Renaming Unit/Free List/Area': 0.0340654,
'Renaming Unit/Free List/Gate Leakage': 2.5481e-05,
'Renaming Unit/Free List/Peak Dynamic': 0.0306032,
'Renaming Unit/Free List/Runtime Dynamic': 0.00321756,
'Renaming Unit/Free List/Subthreshold Leakage': 0.000370144,
'Renaming Unit/Free List/Subthreshold Leakage with power gating': 0.000201064,
'Renaming Unit/Gate Leakage': 0.00708398,
'Renaming Unit/Int Front End RAT/Area': 0.0941223,
'Renaming Unit/Int Front End RAT/Gate Leakage': 0.000283242,
'Renaming Unit/Int Front End RAT/Peak Dynamic': 0.731965,
'Renaming Unit/Int Front End RAT/Runtime Dynamic': 0.0264287,
'Renaming Unit/Int Front End RAT/Subthreshold Leakage': 0.00435488,
'Renaming Unit/Int Front End RAT/Subthreshold Leakage with power gating': 0.00248228,
'Renaming Unit/Peak Dynamic': 3.58947,
'Renaming Unit/Runtime Dynamic': 0.0950671,
'Renaming Unit/Subthreshold Leakage': 0.0552466,
'Renaming Unit/Subthreshold Leakage with power gating': 0.0276461,
'Runtime Dynamic': 1.70473,
'Subthreshold Leakage': 6.16288,
'Subthreshold Leakage with power gating': 2.55328}],
'DRAM': {'Area': 0,
'Gate Leakage': 0,
'Peak Dynamic': 8.59419502088197,
'Runtime Dynamic': 8.59419502088197,
'Subthreshold Leakage': 4.252,
'Subthreshold Leakage with power gating': 4.252},
'L3': [{'Area': 61.9075,
'Gate Leakage': 0.0484137,
'Peak Dynamic': 0.45468,
'Runtime Dynamic': 0.179204,
'Subthreshold Leakage': 6.80085,
'Subthreshold Leakage with power gating': 3.32364}],
'Processor': {'Area': 191.908,
'Gate Leakage': 1.53485,
'Peak Dynamic': 60.4483,
'Peak Power': 93.5606,
'Runtime Dynamic': 9.52364,
'Subthreshold Leakage': 31.5774,
'Subthreshold Leakage with power gating': 13.9484,
'Total Cores/Area': 128.669,
'Total Cores/Gate Leakage': 1.4798,
'Total Cores/Peak Dynamic': 59.9937,
'Total Cores/Runtime Dynamic': 9.34443,
'Total Cores/Subthreshold Leakage': 24.7074,
'Total Cores/Subthreshold Leakage with power gating': 10.2429,
'Total L3s/Area': 61.9075,
'Total L3s/Gate Leakage': 0.0484137,
'Total L3s/Peak Dynamic': 0.45468,
'Total L3s/Runtime Dynamic': 0.179204,
'Total L3s/Subthreshold Leakage': 6.80085,
'Total L3s/Subthreshold Leakage with power gating': 3.32364,
'Total Leakage': 33.1122,
'Total NoCs/Area': 1.33155,
'Total NoCs/Gate Leakage': 0.00662954,
'Total NoCs/Peak Dynamic': 0.0,
'Total NoCs/Runtime Dynamic': 0.0,
'Total NoCs/Subthreshold Leakage': 0.0691322,
'Total NoCs/Subthreshold Leakage with power gating': 0.0259246}} | 75.126915 | 124 | 0.682157 | 8,095 | 68,666 | 5.780482 | 0.065967 | 0.123437 | 0.112837 | 0.093347 | 0.942406 | 0.934541 | 0.922168 | 0.894065 | 0.871861 | 0.854401 | 0 | 0.132389 | 0.224143 | 68,666 | 914 | 125 | 75.126915 | 0.745941 | 0 | 0 | 0.66849 | 0 | 0 | 0.656866 | 0.048058 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | false | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
d545220585c8ef43745228e1c977b6fe55a30efe | 2,829 | py | Python | tests/unit/test_source/test_tsv_source.py | STARInformatics/kgx | a4b4dd73f866486466c03579f62e0c527ef1af1e | [
"BSD-3-Clause"
] | null | null | null | tests/unit/test_source/test_tsv_source.py | STARInformatics/kgx | a4b4dd73f866486466c03579f62e0c527ef1af1e | [
"BSD-3-Clause"
] | null | null | null | tests/unit/test_source/test_tsv_source.py | STARInformatics/kgx | a4b4dd73f866486466c03579f62e0c527ef1af1e | [
"BSD-3-Clause"
] | null | null | null | import os
import pprint
import pytest
from kgx.source import TsvSource
from tests import RESOURCE_DIR
def test_read_tsv():
"""
Read a TSV using TsvSource.
"""
s = TsvSource()
g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test_nodes.tsv'), format='tsv')
nodes = []
for rec in g:
if rec:
nodes.append(rec)
assert len(nodes) == 3
nodes.sort()
n1 = nodes.pop()[-1]
assert n1['id'] == 'CURIE:456'
assert n1['name'] == 'Disease 456'
assert 'biolink:Disease' in n1['category']
assert n1['description'] == '"Node of type Disease, CURIE:456"'
g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test_edges.tsv'), format='tsv')
edges = []
for rec in g:
if rec:
edges.append(rec)
e1 = edges.pop()[-1]
assert 'id' in e1
assert e1['subject'] == 'CURIE:123'
assert e1['object'] == 'CURIE:456'
assert e1['predicate'] == 'biolink:related_to'
assert e1['relation'] == 'biolink:related_to'
assert 'PMID:1' in e1['publications']
def test_read_csv():
"""
Read a CSV using TsvSource.
"""
s = TsvSource()
g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test_nodes.csv'), format='csv')
nodes = []
for rec in g:
if rec:
nodes.append(rec)
assert len(nodes) == 3
nodes.sort()
n1 = nodes.pop()[-1]
assert n1['id'] == 'CURIE:456'
assert n1['name'] == 'Disease 456'
assert 'biolink:Disease' in n1['category']
assert n1['description'] == 'Node of type Disease, CURIE:456'
g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test_edges.csv'), format='csv')
edges = []
for rec in g:
if rec:
print(rec)
edges.append(rec)
e1 = edges.pop()[-1]
assert 'id' in e1
assert e1['subject'] == 'CURIE:123'
assert e1['object'] == 'CURIE:456'
assert e1['predicate'] == 'biolink:related_to'
assert e1['relation'] == 'biolink:related_to'
assert 'PMID:1' in e1['publications']
def test_read_tsv_compressed():
"""
Read a compressed TSV archive using TsvSource.
"""
s = TsvSource()
g = s.parse(filename=os.path.join(RESOURCE_DIR, 'test.tar'), format='tsv', compression='tar')
nodes = []
edges = []
for rec in g:
if rec:
if len(rec) == 4:
edges.append(rec)
else:
nodes.append(nodes)
assert len(nodes) == 3
assert len(edges) == 1
g = s.parse(
filename=os.path.join(RESOURCE_DIR, 'test.tar.gz'), format='tsv', compression='tar.gz'
)
nodes = []
edges = []
for rec in g:
if rec:
if len(rec) == 4:
edges.append(rec)
else:
nodes.append(nodes)
assert len(nodes) == 3
assert len(edges) == 1
| 26.942857 | 97 | 0.563097 | 378 | 2,829 | 4.156085 | 0.179894 | 0.040738 | 0.026735 | 0.057288 | 0.8345 | 0.8345 | 0.8345 | 0.812222 | 0.812222 | 0.812222 | 0 | 0.033382 | 0.279958 | 2,829 | 104 | 98 | 27.201923 | 0.73785 | 0.036055 | 0 | 0.77381 | 0 | 0 | 0.18426 | 0 | 0 | 0 | 0 | 0 | 0.309524 | 1 | 0.035714 | false | 0 | 0.059524 | 0 | 0.095238 | 0.02381 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
638c3d20dddafd292eba1978584a5a3502ee86ef | 24,461 | py | Python | tmapi/tests/indices/test_scoped_index.py | ajenhl/django-tmapi | 02f009e1b508218cf330ca7748c3a1dd110f3e8d | [
"Apache-2.0"
] | 2 | 2015-03-22T03:23:36.000Z | 2017-01-08T10:57:18.000Z | tmapi/tests/indices/test_scoped_index.py | ajenhl/django-tmapi | 02f009e1b508218cf330ca7748c3a1dd110f3e8d | [
"Apache-2.0"
] | null | null | null | tmapi/tests/indices/test_scoped_index.py | ajenhl/django-tmapi | 02f009e1b508218cf330ca7748c3a1dd110f3e8d | [
"Apache-2.0"
] | 1 | 2020-12-28T04:40:34.000Z | 2020-12-28T04:40:34.000Z | # Copyright 2011 Jamie Norrish (jamie@artefact.org.nz)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Module containing tests against the `ScopedIndex` interface.
Most if not all of these tests are ported from the public domain tests
that come with the TMAPI 2.0 distribution (http://www.tmapi.org/2.0/).
"""
from tmapi.exceptions import IllegalArgumentException
from tmapi.indices.scoped_index import ScopedIndex
from tmapi.tests.models.tmapi_test_case import TMAPITestCase
class ScopedIndexTest (TMAPITestCase):
def setUp (self):
super(ScopedIndexTest, self).setUp()
self._index = self.tm.get_index(ScopedIndex)
self._index.open()
def tearDown (self):
super(ScopedIndexTest, self).tearDown()
self._index.close()
def _update_index (self):
if not self._index.is_auto_updated():
self._index.reindex()
def test_association (self):
theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_associations().count())
self.assertEqual(0, self._index.get_associations(theme).count())
self.assertEqual(0, self._index.get_association_themes().count())
scoped = self.create_association()
self.assertEqual(0, scoped.get_scope().count())
self._update_index()
self.assertEqual(1, self._index.get_associations().count())
self.assertTrue(scoped in self._index.get_associations())
self.assertFalse(theme in self._index.get_association_themes())
scoped.add_theme(theme)
self._update_index()
self.assertEqual(0, self._index.get_associations().count())
self.assertFalse(scoped in self._index.get_associations())
self.assertNotEqual(0, self._index.get_association_themes().count())
self.assertEqual(1, self._index.get_association_themes().count())
self.assertTrue(scoped in self._index.get_associations(theme))
self.assertTrue(theme in self._index.get_association_themes())
scoped.remove()
self._update_index()
self.assertEqual(0, self._index.get_associations().count())
self.assertFalse(scoped in self._index.get_associations())
self.assertFalse(theme in self._index.get_association_themes())
def test_association_match_all (self):
theme = self.create_topic()
theme2 = self.create_topic()
unused_theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_associations().count())
self.assertEqual(0, self._index.get_associations(theme).count())
self.assertEqual(0, self._index.get_association_themes().count())
scoped = self.create_association()
self.assertEqual(0, scoped.get_scope().count())
self._update_index()
self.assertEqual(1, self._index.get_associations().count())
self.assertTrue(scoped in self._index.get_associations())
self.assertFalse(theme in self._index.get_association_themes())
scoped.add_theme(theme)
self._update_index()
self.assertEqual(1, self._index.get_association_themes().count())
self.assertTrue(scoped in self._index.get_associations([theme], True))
self.assertTrue(scoped in self._index.get_associations([theme], False))
scoped.add_theme(theme2)
self._update_index()
self.assertEqual(2, self._index.get_association_themes().count())
self.assertTrue(scoped in self._index.get_associations([theme], True))
self.assertTrue(scoped in self._index.get_associations([theme], False))
self.assertTrue(scoped in self._index.get_associations([theme2], True))
self.assertTrue(scoped in self._index.get_associations([theme2], False))
self.assertTrue(scoped in self._index.get_associations(
[theme, theme2], False))
self.assertTrue(scoped in self._index.get_associations(
[theme, theme2], True))
self.assertTrue(scoped in self._index.get_associations(
[theme, unused_theme], False))
self.assertTrue(scoped in self._index.get_associations(
[theme2, unused_theme], False))
self.assertFalse(scoped in self._index.get_associations(
[theme, unused_theme], True))
self.assertFalse(scoped in self._index.get_associations(
[theme2, unused_theme], True))
def test_association_match_all_illegal (self):
self.assertRaises(IllegalArgumentException,
self._index.get_associations, None, True)
def test_occurrence (self):
theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_occurrences().count())
self.assertEqual(0, self._index.get_occurrences(theme).count())
self.assertEqual(0, self._index.get_occurrence_themes().count())
scoped = self.create_occurrence()
self.assertEqual(0, scoped.get_scope().count())
self._update_index()
self.assertEqual(1, self._index.get_occurrences().count())
self.assertTrue(scoped in self._index.get_occurrences())
self.assertFalse(theme in self._index.get_occurrence_themes())
scoped.add_theme(theme)
self._update_index()
self.assertEqual(0, self._index.get_occurrences().count())
self.assertFalse(scoped in self._index.get_occurrences())
self.assertNotEqual(0, self._index.get_occurrence_themes().count())
self.assertEqual(1, self._index.get_occurrence_themes().count())
self.assertTrue(scoped in self._index.get_occurrences(theme))
self.assertTrue(theme in self._index.get_occurrence_themes())
scoped.remove()
self._update_index()
self.assertEqual(0, self._index.get_occurrences().count())
self.assertFalse(scoped in self._index.get_occurrences())
self.assertFalse(theme in self._index.get_occurrence_themes())
def test_occurrence_match_all (self):
theme = self.create_topic()
theme2 = self.create_topic()
unused_theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_occurrences().count())
self.assertEqual(0, self._index.get_occurrences(theme).count())
self.assertEqual(0, self._index.get_occurrence_themes().count())
scoped = self.create_occurrence()
self.assertEqual(0, scoped.get_scope().count())
self._update_index()
self.assertEqual(1, self._index.get_occurrences().count())
self.assertTrue(scoped in self._index.get_occurrences())
self.assertFalse(theme in self._index.get_occurrence_themes())
scoped.add_theme(theme)
self._update_index()
self.assertEqual(1, self._index.get_occurrence_themes().count())
self.assertTrue(scoped in self._index.get_occurrences([theme], True))
self.assertTrue(scoped in self._index.get_occurrences([theme], False))
scoped.add_theme(theme2)
self._update_index()
self.assertEqual(2, self._index.get_occurrence_themes().count())
self.assertTrue(scoped in self._index.get_occurrences([theme], True))
self.assertTrue(scoped in self._index.get_occurrences([theme], False))
self.assertTrue(scoped in self._index.get_occurrences([theme2], True))
self.assertTrue(scoped in self._index.get_occurrences([theme2], False))
self.assertTrue(scoped in self._index.get_occurrences(
[theme, theme2], False))
self.assertTrue(scoped in self._index.get_occurrences(
[theme, theme2], True))
self.assertTrue(scoped in self._index.get_occurrences(
[theme, unused_theme], False))
self.assertTrue(scoped in self._index.get_occurrences(
[theme2, unused_theme], False))
self.assertFalse(scoped in self._index.get_occurrences(
[theme, unused_theme], True))
self.assertFalse(scoped in self._index.get_occurrences(
[theme2, unused_theme], True))
def test_occurrence_match_all_illegal (self):
self.assertRaises(IllegalArgumentException, self._index.get_occurrences,
None, True)
def test_name (self):
theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_names().count())
self.assertEqual(0, self._index.get_names(theme).count())
self.assertEqual(0, self._index.get_name_themes().count())
scoped = self.create_name()
self.assertEqual(0, scoped.get_scope().count())
self._update_index()
self.assertEqual(1, self._index.get_names().count())
self.assertTrue(scoped in self._index.get_names())
self.assertFalse(theme in self._index.get_name_themes())
scoped.add_theme(theme)
self._update_index()
self.assertEqual(0, self._index.get_names().count())
self.assertFalse(scoped in self._index.get_names())
self.assertNotEqual(0, self._index.get_name_themes().count())
self.assertEqual(1, self._index.get_name_themes().count())
self.assertTrue(scoped in self._index.get_names(theme))
self.assertTrue(theme in self._index.get_name_themes())
scoped.remove()
self._update_index()
self.assertEqual(0, self._index.get_names().count())
self.assertFalse(scoped in self._index.get_names())
self.assertFalse(theme in self._index.get_name_themes())
def test_name_2 (self):
theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_names().count())
self.assertEqual(0, self._index.get_names(theme).count())
self.assertEqual(0, self._index.get_name_themes().count())
scoped = self.create_topic().create_name('tinyTiM', scope=theme)
self.assertEqual(1, scoped.get_scope().count())
self._update_index()
self.assertEqual(0, self._index.get_names().count())
self.assertFalse(scoped in self._index.get_names())
self.assertNotEqual(0, self._index.get_name_themes().count())
self.assertEqual(1, self._index.get_name_themes().count())
self.assertTrue(scoped in self._index.get_names(theme))
self.assertTrue(theme in self._index.get_name_themes())
scoped.remove()
self._update_index()
self.assertEqual(0, self._index.get_names().count())
self.assertFalse(scoped in self._index.get_names())
self.assertEqual(0, self._index.get_names(theme).count())
self.assertFalse(theme in self._index.get_name_themes())
def test_name_match_all (self):
theme = self.create_topic()
theme2 = self.create_topic()
unused_theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_names().count())
self.assertEqual(0, self._index.get_names(theme).count())
self.assertEqual(0, self._index.get_name_themes().count())
scoped = self.create_name()
self.assertEqual(0, scoped.get_scope().count())
self._update_index()
self.assertEqual(1, self._index.get_names().count())
self.assertTrue(scoped in self._index.get_names())
self.assertFalse(theme in self._index.get_name_themes())
scoped.add_theme(theme)
self._update_index()
self.assertEqual(1, self._index.get_name_themes().count())
self.assertTrue(scoped in self._index.get_names([theme], True))
self.assertTrue(scoped in self._index.get_names([theme], False))
scoped.add_theme(theme2)
self._update_index()
self.assertEqual(2, self._index.get_name_themes().count())
self.assertTrue(scoped in self._index.get_names([theme], True))
self.assertTrue(scoped in self._index.get_names([theme], False))
self.assertTrue(scoped in self._index.get_names([theme2], True))
self.assertTrue(scoped in self._index.get_names([theme2], False))
self.assertTrue(scoped in self._index.get_names([theme, theme2], True))
self.assertTrue(scoped in self._index.get_names([theme, theme2], False))
self.assertTrue(scoped in self._index.get_names(
[theme, unused_theme], False))
self.assertTrue(scoped in self._index.get_names(
[theme2, unused_theme], False))
self.assertFalse(scoped in self._index.get_names(
[theme, unused_theme], True))
self.assertFalse(scoped in self._index.get_names(
[theme2, unused_theme], True))
def test_name_match_all_illegal (self):
self.assertRaises(IllegalArgumentException, self._index.get_names,
None, True)
def test_variant_illegal (self):
self.assertRaises(IllegalArgumentException, self._index.get_variants,
None)
def test_variant_match_all_illegal (self):
self.assertRaises(IllegalArgumentException, self._index.get_variants,
None, True)
def test_variant (self):
theme = self.create_topic()
theme2 = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_variants(theme).count())
self.assertEqual(0, self._index.get_variant_themes().count())
name = self.create_name()
self.assertEqual(0, name.get_scope().count())
scoped = name.create_variant('Variant', theme)
self.assertEqual(1, scoped.get_scope().count(),
'Unexpected variant\'s scope size')
self._update_index()
self.assertNotEqual(0, self._index.get_variant_themes().count())
self.assertEqual(1, self._index.get_variant_themes().count(),
'Unexpected number of variant themes')
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
name.add_theme(theme2)
self.assertEqual(1, name.get_scope().count())
self.assertEqual(2, scoped.get_scope().count(), 'The scope change of the parent is not reflected in the variant\'s scope')
self._update_index()
self.assertEqual(2, self._index.get_variant_themes().count(), 'Change of the parent\'s scope is not reflected in the index')
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
self.assertTrue(scoped in self._index.get_variants(theme2))
self.assertTrue(theme2 in self._index.get_variant_themes())
name.remove_theme(theme2)
self._update_index()
self.assertNotEqual(0, self._index.get_variant_themes().count())
self.assertEqual(1, self._index.get_variant_themes().count(), 'The scope change in the name is not reflected in the variant')
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
scoped.add_theme(theme2)
self._update_index()
self.assertEqual(2, self._index.get_variant_themes().count(), 'Change of the variant\'s scope is not reflected in the index')
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
self.assertTrue(scoped in self._index.get_variants(theme2))
self.assertTrue(theme2 in self._index.get_variant_themes())
name.add_theme(theme2)
self._update_index()
self.assertEqual(2, self._index.get_variant_themes().count(), 'Adding a theme to the variant\'s parent is not reflected in the index')
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
self.assertTrue(scoped in self._index.get_variants(theme2))
self.assertTrue(theme2 in self._index.get_variant_themes())
name.remove_theme(theme2)
self._update_index()
self.assertEqual(2, self._index.get_variant_themes().count(), 'Removing the name\'s theme MUST NOT be reflected in the variant\'s scope')
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
self.assertTrue(scoped in self._index.get_variants(theme2))
self.assertTrue(theme2 in self._index.get_variant_themes())
scoped.remove_theme(theme2)
self.assertNotEqual(0, self._index.get_variant_themes().count())
self.assertEqual(1, self._index.get_variant_themes().count())
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
def test_variant_2 (self):
theme = self.create_topic()
theme2 = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_variants(theme).count())
self.assertEqual(0, self._index.get_variants(theme2).count())
self.assertEqual(0, self._index.get_variant_themes().count())
name = self.create_topic().create_name('Name', scope=theme2)
self.assertEqual(1, name.get_scope().count())
scoped = name.create_variant('Variant', theme)
self.assertEqual(2, scoped.get_scope().count())
self._update_index()
self.assertEqual(2, self._index.get_variant_themes().count())
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
self.assertTrue(scoped in self._index.get_variants(theme2))
self.assertTrue(theme2 in self._index.get_variant_themes())
name.remove_theme(theme2)
self.assertEqual(0, name.get_scope().count())
self._update_index()
self.assertEqual(1, self._index.get_variant_themes().count())
self.assertTrue(scoped in self._index.get_variants(theme))
self.assertTrue(theme in self._index.get_variant_themes())
def test_variant_match_all (self):
theme = self.create_topic()
theme2 = self.create_topic()
unused_theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_variants(theme).count())
self.assertEqual(0, self._index.get_variants(theme2).count())
self.assertEqual(0, self._index.get_variant_themes().count())
name = self.create_topic().create_name('Name')
self.assertEqual(0, name.get_scope().count())
scoped = name.create_variant('Variant', theme)
self.assertEqual(1, scoped.get_scope().count())
self._update_index()
self.assertEqual(1, self._index.get_variant_themes().count())
self.assertTrue(scoped in self._index.get_variants([theme], True))
self.assertTrue(scoped in self._index.get_variants([theme], False))
self.assertFalse(scoped in self._index.get_variants([theme2], True))
self.assertFalse(scoped in self._index.get_variants([theme2], False))
scoped.add_theme(theme2)
self._update_index()
self.assertTrue(scoped in self._index.get_variants([theme], True))
self.assertTrue(scoped in self._index.get_variants([theme], False))
self.assertTrue(scoped in self._index.get_variants([theme2], True))
self.assertTrue(scoped in self._index.get_variants([theme2], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2], True))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2, unused_theme], False))
self.assertFalse(scoped in self._index.get_variants(
[theme, theme2, unused_theme], True))
name_theme = self.create_topic()
name.add_theme(name_theme)
self._update_index()
self.assertTrue(scoped in self._index.get_variants([theme], True))
self.assertTrue(scoped in self._index.get_variants([theme], False))
self.assertTrue(scoped in self._index.get_variants([theme2], True))
self.assertTrue(scoped in self._index.get_variants([theme2], False))
self.assertTrue(scoped in self._index.get_variants([name_theme], True))
self.assertTrue(scoped in self._index.get_variants([name_theme], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2], True))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2, name_theme], True))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2, name_theme], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2, unused_theme], False))
self.assertFalse(scoped in self._index.get_variants(
[theme, theme2, unused_theme], True))
name.remove_theme(name_theme)
self._update_index()
self.assertFalse(scoped in self._index.get_variants([name_theme], True))
self.assertFalse(scoped in self._index.get_variants(
[name_theme], False))
self.assertFalse(scoped in self._index.get_variants(
[theme, theme2, name_theme], True))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2, name_theme], False))
scoped.remove_theme(theme)
self._update_index()
self.assertFalse(scoped in self._index.get_variants([theme], True))
self.assertFalse(scoped in self._index.get_variants([theme], False))
self.assertFalse(scoped in self._index.get_variants(
[theme, theme2], True))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2], False))
def test_variant_match_all_2 (self):
theme = self.create_topic()
theme2 = self.create_topic()
unused_theme = self.create_topic()
name_theme = self.create_topic()
self._update_index()
self.assertEqual(0, self._index.get_variants(theme).count())
self.assertEqual(0, self._index.get_variants(theme2).count())
self.assertEqual(0, self._index.get_variant_themes().count())
name = self.create_topic().create_name('Name', scope=name_theme)
self.assertEqual(1, name.get_scope().count())
scoped = name.create_variant('Variant', [theme, theme2])
self.assertEqual(3, scoped.get_scope().count())
self._update_index()
self.assertEqual(3, self._index.get_variant_themes().count())
self.assertTrue(scoped in self._index.get_variants([theme], True))
self.assertTrue(scoped in self._index.get_variants([theme], False))
self.assertTrue(scoped in self._index.get_variants([theme2], True))
self.assertTrue(scoped in self._index.get_variants([theme2], False))
self.assertTrue(scoped in self._index.get_variants([name_theme], True))
self.assertTrue(scoped in self._index.get_variants([name_theme], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2], True))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2, name_theme], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2, name_theme], False))
self.assertTrue(scoped in self._index.get_variants(
[theme, theme2, unused_theme], False))
self.assertFalse(scoped in self._index.get_variants(
[theme, theme2, unused_theme], True))
| 52.717672 | 145 | 0.673685 | 3,022 | 24,461 | 5.196889 | 0.049967 | 0.12894 | 0.168099 | 0.124801 | 0.915123 | 0.903725 | 0.896721 | 0.885132 | 0.875963 | 0.859472 | 0 | 0.009638 | 0.206778 | 24,461 | 463 | 146 | 52.831533 | 0.799773 | 0.031847 | 0 | 0.805226 | 0 | 0 | 0.013608 | 0 | 0 | 0 | 0 | 0 | 0.565321 | 1 | 0.045131 | false | 0 | 0.007126 | 0 | 0.054632 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 |
98460f90297c5bb37bad032f1201cc32789bbc1a | 56,611 | py | Python | adapters/adapter.py | kweveen/tng-vnv-platform-adapter | 90b8499bb564664a751af26ff20f7ddbd5358b6a | [
"Apache-2.0"
] | null | null | null | adapters/adapter.py | kweveen/tng-vnv-platform-adapter | 90b8499bb564664a751af26ff20f7ddbd5358b6a | [
"Apache-2.0"
] | null | null | null | adapters/adapter.py | kweveen/tng-vnv-platform-adapter | 90b8499bb564664a751af26ff20f7ddbd5358b6a | [
"Apache-2.0"
] | null | null | null | #!/usr/bin/python
from flask import Flask, request, jsonify, render_template
import os, sys, logging, uuid, json
from werkzeug import secure_filename
#import serviceplatform
import psycopg2
import requests
import subprocess
import models.database as database
import re
import ast
from ast import literal_eval
FILE = "db-config.cfg"
class Adapter:
def __init__(self, name):
self.name = name
self.host = "host"
self.type = "type"
def getName(self):
return self.name
def setName(self, newName):
self.name = newName
def getHost(self):
return self.host
def setHost(self, newHost):
self.host = newHost
def getType(self):
return self.type
def setType(self, newType):
self.type = newType
def updateToken(self,token):
try:
db = database.Database(FILE)
connection = psycopg2.connect(user = db.user,
password = db.password,
host = db.host,
port = db.port,
database = db.database)
cursor = connection.cursor()
print ( connection.get_dsn_parameters(),"\n")
print (self.name)
get_type = "SELECT type FROM service_platforms WHERE name=\'" +self.name+ "\'"
print (get_type)
update_token = "UPDATE service_platforms SET service_token = \'" +token+ "\' WHERE name = \'" +self.name+ "\'"
print (update_token)
cursor.execute(update_token)
connection.commit()
return "token updated", 200
except (Exception, psycopg2.Error) as error :
print (error)
exception_message = str(error)
return exception_message, 401
finally:
#closing database connection.
if(connection):
cursor.close()
connection.close()
print("PostgreSQL connection is closed")
def getDBType(self):
try:
db = database.Database(FILE)
connection = psycopg2.connect(user = db.user,
password = db.password,
host = db.host,
port = db.port,
database = db.database)
cursor = connection.cursor()
print ( connection.get_dsn_parameters(),"\n")
#create table Service Platforms
get_type = "SELECT type FROM service_platforms WHERE name=\'" +self.name+ "\'"
print (get_type)
cursor.execute(get_type)
all = cursor.fetchall()
#return jsonify(all), 200
type_0 = all.__str__()
print(type_0)
type_1 = type_0[3:]
print(type_1)
type_2 = type_1[:-4]
print(type_2)
return type_2
except (Exception, psycopg2.Error) as error :
print (error)
exception_message = str(error)
return exception_message, 401
finally:
#closing database connection.
if(connection):
cursor.close()
connection.close()
print("PostgreSQL connection is closed")
def getDBUserName(self):
try:
db = database.Database(FILE)
connection = psycopg2.connect(user = db.user,
password = db.password,
host = db.host,
port = db.port,
database = db.database)
cursor = connection.cursor()
print ( connection.get_dsn_parameters(),"\n")
#create table Service Platforms
get_username = "SELECT username FROM service_platforms WHERE name=\'" +self.name+ "\'"
print (get_username)
cursor.execute(get_username)
all = cursor.fetchall()
#return jsonify(all), 200
type_0 = all.__str__()
print(type_0)
type_1 = type_0[3:]
print(type_1)
type_2 = type_1[:-4]
print(type_2)
return type_2
except (Exception, psycopg2.Error) as error :
print (error)
exception_message = str(error)
return exception_message, 401
finally:
#closing database connection.
if(connection):
cursor.close()
connection.close()
print("PostgreSQL connection is closed")
def getDBProjectName(self):
try:
db = database.Database(FILE)
connection = psycopg2.connect(user = db.user,
password = db.password,
host = db.host,
port = db.port,
database = db.database)
cursor = connection.cursor()
print ( connection.get_dsn_parameters(),"\n")
#create table Service Platforms
get_project_name = "SELECT project_name FROM service_platforms WHERE name=\'" +self.name+ "\'"
print (get_project_name)
cursor.execute(get_project_name)
all = cursor.fetchall()
#return jsonify(all), 200
type_0 = all.__str__()
print(type_0)
type_1 = type_0[3:]
print(type_1)
type_2 = type_1[:-4]
print(type_2)
return type_2
except (Exception, psycopg2.Error) as error :
print (error)
exception_message = str(error)
return exception_message, 401
finally:
#closing database connection.
if(connection):
cursor.close()
connection.close()
print("PostgreSQL connection is closed")
def getDBPassword(self):
try:
db = database.Database(FILE)
connection = psycopg2.connect(user = db.user,
password = db.password,
host = db.host,
port = db.port,
database = db.database)
cursor = connection.cursor()
print ( connection.get_dsn_parameters(),"\n")
#create table Service Platforms
get_password= "SELECT username FROM service_platforms WHERE name=\'" +self.name+ "\'"
print (get_password)
cursor.execute(get_password)
all = cursor.fetchall()
#return jsonify(all), 200
type_0 = all.__str__()
print(type_0)
type_1 = type_0[3:]
print(type_1)
type_2 = type_1[:-4]
print(type_2)
return type_2
except (Exception, psycopg2.Error) as error :
print (error)
exception_message = str(error)
return exception_message, 401
finally:
#closing database connection.
if(connection):
cursor.close()
connection.close()
print("PostgreSQL connection is closed")
def getDBProject(self):
try:
db = database.Database(FILE)
connection = psycopg2.connect(user = db.user,
password = db.password,
host = db.host,
port = db.port,
database = db.database)
cursor = connection.cursor()
print ( connection.get_dsn_parameters(),"\n")
#create table Service Platforms
get_password= "SELECT project_name FROM service_platforms WHERE name=\'" +self.name+ "\'"
print (get_password)
cursor.execute(get_password)
all = cursor.fetchall()
#return jsonify(all), 200
type_0 = all.__str__()
print(type_0)
type_1 = type_0[3:]
print(type_1)
type_2 = type_1[:-4]
print(type_2)
return type_2
except (Exception, psycopg2.Error) as error :
print (error)
exception_message = str(error)
return exception_message, 401
finally:
#closing database connection.
if(connection):
cursor.close()
connection.close()
print("PostgreSQL connection is closed")
def getDBHost(self):
try:
db = database.Database(FILE)
connection = psycopg2.connect(user = db.user,
password = db.password,
host = db.host,
port = db.port,
database = db.database)
cursor = connection.cursor()
print ( connection.get_dsn_parameters(),"\n")
#create table Service Platforms
print (self.name)
get_host = "SELECT host FROM service_platforms WHERE name=\'" +self.name+ "\'"
print (get_host)
cursor.execute(get_host)
all = cursor.fetchall()
#return jsonify(all), 200
return all, 200
except (Exception, psycopg2.Error) as error :
print (error)
exception_message = str(error)
return exception_message, 401
finally:
#closing database connection.
if(connection):
cursor.close()
connection.close()
print("PostgreSQL connection is closed")
def getPackages(self):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/packages'
#url = sp_url + '/packages'
response = requests.get(url, headers=JSON_CONTENT_HEADER)
if response.ok:
return (response.text, response.status_code, response.headers.items())
if my_type == 'osm':
return "osm packages"
def getPackage(self,name,vendor,version):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/packages'
print (name,vendor,version)
response = requests.get(url,headers=JSON_CONTENT_HEADER)
response_json = response.content
jjson = json.loads(response_json)
pkg = [x for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version]
if response.ok:
print(pkg)
return jsonify(pkg)
def deletePackage(self,name,vendor,version):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/packages'
print (name,vendor,version)
response = requests.get(url,headers=JSON_CONTENT_HEADER)
response_json = response.content
jjson = json.loads(response_json)
pkg = [x for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version]
if pkg:
print(pkg)
#uuid_to_delete = pkg['pd']['uuid']
#uuid_to_delete_1 = [uuid for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version]
uuid_to_delete_1 = [obj['uuid'] for obj in jjson if(obj['pd']['name'] == name)]
print(uuid_to_delete_1)
uuid_0 = uuid_to_delete_1.__str__()
uuid_to_delete_2 = uuid_0[2:]
print(uuid_to_delete_2)
uuid_to_delete_3 = uuid_to_delete_2[:-2]
print(uuid_to_delete_3)
url_for_delete = url + '/' + uuid_to_delete_3
print (url_for_delete)
delete = requests.delete(url_for_delete, headers=JSON_CONTENT_HEADER)
if response.ok:
return (delete.text, delete.status_code, delete.headers.items())
def getPackagebyId(self,name,vendor,version):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/packages'
print (name,vendor,version)
response = requests.get(url,headers=JSON_CONTENT_HEADER)
response_json = response.content
jjson = json.loads(response_json)
pkg = [x for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version]
if pkg:
print(pkg)
#uuid_to_delete = pkg['pd']['uuid']
#uuid_to_delete_1 = [uuid for x in jjson if x['pd']['name'] == name and x['pd']['vendor'] == vendor and x['pd']['version'] == version]
uuid_to_delete_1 = [obj['uuid'] for obj in jjson if(obj['pd']['name'] == name)]
print(uuid_to_delete_1)
uuid_0 = uuid_to_delete_1.__str__()
uuid_to_delete_2 = uuid_0[2:]
print(uuid_to_delete_2)
uuid_to_delete_3 = uuid_to_delete_2[:-2]
print(uuid_to_delete_3)
url_for_delete = url + '/' + uuid_to_delete_3
print (url_for_delete)
delete = requests.get(url_for_delete, headers=JSON_CONTENT_HEADER)
if response.ok:
return (delete.text, delete.status_code, delete.headers.items())
def uploadPackage(self,package):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/packages'
print(package)
print(url)
files = {'package': open(package,'rb')}
upload = requests.post(url, files=files)
if request.method == 'POST':
return upload.text
#curl -v -i -X POST -F "package=@./eu.5gtango.test-connectivity.0.1.tgo" http://sta-vnv-ave-v4-0.5gtango.eu:32002/api/v3/packages
#upload_nsd_3 = upload_nsd_2 + " --data-binary "
#upload_nsd_4 = upload_nsd_3 + "\"@" +file_to_upload+ "\" " + url_2
#print (upload_nsd_4)
#upload = subprocess.check_output([upload_nsd_4], shell=True)
#return jsonify(upload_nsd_4)
#return (upload)
#upload_package_1="curl -v -i -X POST -F \"package=@." + package + "\" " + url
#return upload_package_1
def uploadOSMService(self,request):
#JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'osm':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
content = request.get_json()
print ("request content:")
print (content)
token = self.getOSMToken(request)
print (token)
file_to_upload = content['service']
print ("File to upload")
print (file_to_upload)
file_composed = "@" + file_to_upload
print (file_composed)
#file = {'file': open(file_to_upload, 'rb')}
file = {'nsd-create': open(file_to_upload, 'rb')}
print (file)
data = {'service':file_to_upload}
#data = request.get_json()
print (data)
HEADERS = {
'Accept':'application/yaml',
'Content-Type':'application/zip',
'Authorization':'Bearer ' +token+''
}
print (HEADERS)
url = sp_host_2 + ':9999/osm/nsd/v1/ns_descriptors_content'
url_2 = url.replace("http","https")
print (url_2)
#upload_nsddddd = "osm --hostname " + sp_host_3 + " vnfd-create " + url_2
upload_nsd = "curl -X POST --insecure -w \"%{http_code}\" -H \"Content-type: application/zip\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer "
upload_nsd_2 = upload_nsd +token + "\" "
upload_nsd_3 = upload_nsd_2 + " --data-binary "
upload_nsd_4 = upload_nsd_3 + "\"@" +file_to_upload+ "\" " + url_2
print (upload_nsd_4)
upload = subprocess.check_output([upload_nsd_4], shell=True)
#return jsonify(upload_nsd_4)
return (upload)
def uploadOSMFunction(self,request):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'osm':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
token = self.getOSMToken(request)
print (token)
content = request.get_json()
file_to_upload = content['function']
url = sp_host_2 + ':9999/osm/vnfpkgm/v1/vnf_packages_content'
url_2 = url.replace("http","https")
print (url_2)
upload_nsd = "curl -X POST --insecure -w \"%{http_code}\" -H \"Content-type: application/zip\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer "
upload_nsd_2 = upload_nsd +token + "\" "
upload_nsd_3 = upload_nsd_2 + " --data-binary "
upload_nsd_4 = upload_nsd_3 + "\"@" +file_to_upload+ "\" " + url_2
print (upload_nsd_4)
upload = subprocess.check_output([upload_nsd_4], shell=True)
#return jsonify(upload_nsd_4)
return (upload)
def getServices(self):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
print (my_type)
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/services'
print (url)
#url = sp_url + '/packages'
response = requests.get(url, headers=JSON_CONTENT_HEADER)
if response.ok:
return (response.text, response.status_code, response.headers.items())
if my_type == 'osm':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_2 + ':32002/api/v3/services'
get_nsd_list = "osm --hostname " + sp_host_3 + " nsd-list"
print (get_nsd_list)
#get = os.system(get_nsd_list).__str__()
#return get
get = subprocess.check_output([get_nsd_list], shell=True)
return (get)
def getFunctions(self):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/functions'
#url = sp_url + '/packages'
response = requests.get(url, headers=JSON_CONTENT_HEADER)
if response.ok:
return (response.text, response.status_code, response.headers.items())
if my_type == 'osm':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_2 + ':32002/api/v3/functions'
get_vnfd_list = "osm --hostname " + sp_host_3 + " vnfd-list"
print (get_vnfd_list)
#get = os.system(get_nsd_list).__str__()
#return get
get = subprocess.check_output([get_vnfd_list], shell=True)
return (get)
def getService(self,name,vendor,version):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/services'
print (name,vendor,version)
response = requests.get(url,headers=JSON_CONTENT_HEADER)
response_json = response.content
jjson = json.loads(response_json)
pkg = [x for x in jjson if x['nsd']['name'] == name and x['nsd']['vendor'] == vendor and x['nsd']['version'] == version]
if response.ok:
print(pkg)
return jsonify(pkg)
def getServiceInstantiations(self,name,vendor,version):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/requests'
print (name,vendor,version)
response = requests.get(url,headers=JSON_CONTENT_HEADER)
response_json = response.content
#response_json = response.text
print (response_json)
#json_str = json.dumps(response._content)
#jjson = json.loads(response.text.__str__())
#jjson = json.loads(response_json)
jjson = json.loads(response.content)
print (jjson)
#services = [x for x in jjson if x['service']['name'] == name and x['service']['vendor'] == vendor and x['service']['version'] == version]
#print (jjson['service']['name'])
#print (obj for obj in jjson if(obj['service']['vendor'] == 'eu.5gtango'))
#illo = [obj for obj in jjson if(obj['service'] == '11111')]
print ("illo")
#idd = print ([obj for obj in jjson[0]['service']['uuid']])
idd = print (jjson[0]['service']['uuid'])
idd = print (jjson[0]['service']['name'])
idd = print (jjson[1]['service']['uuid'])
idd = print (jjson[1]['service']['name'])
#idd = [obj.service['name'] for obj in jjson]
print ("illo")
N = 0
for N in range(10000):
print (jjson['service']['uuid'])
N = N + 1
print (N)
if response.ok:
#return (response.text, response.status_code, response.headers.items())
return jsonify("no")
def getServiceId(self,name,vendor,version):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/services'
print (name,vendor,version)
response = requests.get(url,headers=JSON_CONTENT_HEADER)
response_json = response.content
print (response_json)
jjson = json.loads(response_json)
pkg = [x for x in jjson if x['nsd']['name'] == name and x['nsd']['vendor'] == vendor and x['nsd']['version'] == version]
if pkg:
print(pkg)
uuid_to_delete_1 = [obj['uuid'] for obj in jjson if(obj['nsd']['name'] == name)]
print(uuid_to_delete_1)
uuid_0 = uuid_to_delete_1.__str__()
uuid_to_delete_2 = uuid_0[2:]
print(uuid_to_delete_2)
uuid_to_delete_3 = uuid_to_delete_2[:-2]
print(uuid_to_delete_3)
url_for_delete = url + '/' + uuid_to_delete_3
print (url_for_delete)
delete = requests.get(url_for_delete, headers=JSON_CONTENT_HEADER)
if response.ok:
return uuid_to_delete_3
def instantiationStatus(self,request):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
print('this SP is a Sonata')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
#url = sp_host_2 + ':32002/api/v3/requests/' + id
url = sp_host_2 + ':32002/api/v3/requests/' + request
print (url)
response = requests.get(url,headers=JSON_CONTENT_HEADER)
response_json = response.content
print (response_json)
#return response_json
if response.ok:
return (response.text, response.status_code, response.headers.items())
if my_type == 'osm':
print('this SP is a OSM')
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
token = self.getOSMToken(request)
print (token)
content = request.get_json()
ns_id = content['ns_id']
print (ns_id)
url = sp_host_2 + ':9999/osm/nslcm/v1/ns_instances'
url_2 = url.replace("http","https")
print (url_2)
status_ns = "curl --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer "
status_ns_2 = status_ns +token + "\" "
status_ns_3 = status_ns_2 + " " + url_2 + "/" + ns_id
print (status_ns_3)
status = subprocess.check_output([status_ns_3], shell=True)
return (status)
def instantiationsStatus(self):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
print('this SP is a Sonata')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/requests'
print (url)
response = requests.get(url,headers=JSON_CONTENT_HEADER)
response_json = response.content
print (response_json)
#return response_json
if response.ok:
return (response.text, response.status_code, response.headers.items())
#print("status")
#return "status"
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
get_status = "osm --hostname " + sp_host_3 + " ns-list"
print (get_status)
status = subprocess.check_output([get_status], shell=True)
return (status)
def instantiation(self,request):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
JSON_CONTENT_HEADER = {'content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
print('this SP is a Sonata')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/requests'
print (url)
print(request.get_json())
data = request.get_json()
print(url)
print (data)
#upload = requests.post(url, files=files)
#upload = requests.post(url, files=files)
#instantiate = requests.post(url,data,headers=JSON_CONTENT_HEADER)
instantiate = requests.post( url, data=json.dumps(data), headers=JSON_CONTENT_HEADER)
print (instantiate)
if request.method == 'POST':
return instantiate.text
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
print(request.get_json())
content = request.get_json()
print(content)
token = self.getOSMToken(request)
print (token)
content = request.get_json()
print (content)
print (content['nsd_name'])
print (content['ns_name'])
print (content['vim_account'])
url = sp_host_2 + ':9999/osm/nslcm/v1/ns_instances_content'
url_2 = url.replace("http","https")
print (url_2)
print (content['vim_account'])
vim_id = self.getVimId(content['vim_account'])
print (vim_id)
print (content['nsd_name'])
nsd_id = self.getOSMNsdId(content['nsd_name'])
print (nsd_id)
HEADERS = {
'Accept':'application/json',
'Content-Type':'application/json',
'Authorization':'Bearer ' +token+''
}
data_inst = {
'nsdId':''+nsd_id+'',
'nsName':''+content['ns_name']+'',
'vimAccountId':''+vim_id+''
}
instantiate_nsd = "curl -X POST --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer "
instantiate_nsd_2 = instantiate_nsd +token + "\" "
instantiate_nsd_3 = instantiate_nsd_2 + " --data \"" + str(data_inst) + "\""
instantiate_nsd_4 = instantiate_nsd_3 + " " + url_2
print (instantiate_nsd_4)
inst = subprocess.check_output([instantiate_nsd_4], shell=True)
return (inst)
def instantiationDelete(self,request):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
print('this SP is a Sonata')
print('this SP is a Sonata')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/requests'
print(request.get_json())
data = request.get_json()
print(url)
#upload = requests.post(url, files=files)
#upload = requests.post(url, files=files)
terminate = requests.post(url,data,headers=JSON_CONTENT_HEADER)
if request.method == 'POST':
return terminate.text
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
print(request.get_json())
print(url)
token = self.getOSMToken(request)
print (token)
content = request.get_json()
url = sp_host_2 + ':9999/osm/nslcm/v1/ns_instances_content'
url_2 = url.replace("http","https")
print (url_2)
print (content['ns_id'])
ns_id = content['ns_id']
print (ns_id)
delete_ns = "curl -X DELETE --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer "
delete_ns_2 = delete_ns +token + "\" "
delete_ns_3 = delete_ns_2 + " " + url_2 + "/" + ns_id
print (delete_ns_3)
terminate = subprocess.check_output([delete_ns_3], shell=True)
return (terminate)
#delete_ns = "osm --hostname " + sp_host_3 + " ns-delete " + data['ns_name']
#print (delete_ns)
#delete = subprocess.check_output([delete_ns], shell=True)
#return (delete)
def getOSMToken(self,request):
#JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
JSON_CONTENT_HEADER = {'Accept':'application/json'}
my_type = self.getDBType()
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
#url = sp_host_2 + '/requests'
url = sp_host_2 + ':9999/osm/admin/v1/tokens'
url_2 = url.replace("http","https")
print (url_2)
print(request.get_json())
#data = request.get_json()
print(url_2)
#print (data)
#print (data['nsd_name'])
#print (data['username'])
#print (data['password'])
#print (data['project_id'])
pr_name = self.getDBProjectName()
print ("project name from DB:")
print (pr_name)
if pr_name:
project_id_for_token = pr_name
if not pr_name:
data = request.get_json()
project_id_for_token = data['project_id']
print ("project name from json body:")
print (pr_name)
#project_id_for_token = data['project_id']
print (project_id_for_token)
#username_for_token = data['username']
#password_for_token = data['password']
username_for_token = self.getDBUserName()
password_for_token = self.getDBPassword()
#project_id_for_token = data['project_id']
admin_data = "{username: 'admin', password: 'admin', project_id: 'admin'}"
print (admin_data)
#update_token = "UPDATE service_platforms SET service_token = \'" +token+ "\' WHERE name = \'" +self.name+ "\'"
data_for_token= "{username: \'" +username_for_token+ "\', password: \'" +password_for_token+ "\', project_id: \'" +project_id_for_token+ "\'}"
#print (data)
get_token = requests.post(url_2,data=data_for_token,headers=JSON_CONTENT_HEADER,verify=False)
print (get_token.text)
print (get_token.content)
token_id = get_token.json()
print (token_id['id'])
#upd_tok = self.updateToken(token_id['id'])
#print (upd_tok)
return token_id['id']
def getVims(self):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
print('this SP is a Sonata')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/requests'
print (url)
return url
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
get_vims = "osm --hostname " + sp_host_3 + " vim-list"
print (get_vims)
vims = subprocess.check_output([get_vims], shell=True)
return (vims)
def getVim(self,vim):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
print('this SP is a Sonata')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/requests'
print (url)
return url
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
get_vim = "osm --hostname " + sp_host_3 + " vim-show " + vim
print (get_vim)
vim_info = subprocess.check_output([get_vim], shell=True)
return jsonify(vim_info)
def getVimId(self,vim):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
print('this SP is a Sonata')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/requests'
print (url)
return url
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
get_vim = "osm --hostname " + sp_host_3 + " vim-show " + vim
print (get_vim)
vim_info = subprocess.check_output([get_vim], shell=True)
print (vim_info)
print (type(vim_info))
s = json.dumps(str(vim_info))
print(s)
print (type(s))
print ("ILLO")
start = s.find('_id')
end = s.find('\\\" ', start)
print (s[start+20:end])
vim_id = s[start+20:end]
print ("ILLO")
return vim_id
def getOSMNsdId(self,nsd_name):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
get_nsd = "osm --hostname " + sp_host_3 + " nsd-show " + nsd_name
print (get_nsd)
nsd_info = subprocess.check_output([get_nsd], shell=True)
print (nsd_info)
print (type(nsd_info))
s = json.dumps(str(nsd_info))
print(s)
print (type(s))
print ("ILLO")
start = s.find('_id')
end = s.find('\\\" ', start)
#print (s[start+21:end])
#vim_id = s[start+21:end]
print (s[start+21:end])
vim_id = s[start+21:end]
print ("ILLO")
return vim_id
def downloadPackage(self):
JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
my_type = self.getDBType()
if my_type == 'sonata':
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
url = sp_host_2 + ':32002/api/v3/packages'
print(url)
files = {'package': open(package,'rb')}
upload = requests.post(url, files=files)
if request.method == 'POST':
return upload.text
def deleteOSMService(self,id_to_delete):
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
token = self.getOSMTokenForDelete()
print (token)
url = sp_host_2 + ':9999/osm/nsd/v1/ns_descriptors_content'
url_2 = url.replace("http","https")
print (url_2)
delete_nsd = "curl --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer "
delete_nsd_2 = delete_nsd +token + "\" " + url_2 + "/" + id_to_delete + " -X DELETE"
print (delete_nsd_2)
deletion = subprocess.check_output([delete_nsd_2], shell=True)
return (deletion)
def deleteOSMFunction(self,id_to_delete):
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
sp_host_3 = sp_host_2[7:]
print ("sp3 es: ")
print (sp_host_3)
url = sp_host_3
token = self.getOSMTokenForDelete()
print (token)
url = sp_host_2 + ':9999/osm/vnfpkgm/v1/vnf_packages'
url_2 = url.replace("http","https")
print (url_2)
delete_nsd = "curl --insecure -w \"%{http_code}\" -H \"Content-type: application/yaml\" -H \"Accept: application/yaml\" -H \"Authorization: Bearer "
delete_nsd_2 = delete_nsd +token + "\" " + url_2 + "/" + id_to_delete + " -X DELETE"
print (delete_nsd_2)
deletion = subprocess.check_output([delete_nsd_2], shell=True)
return (deletion)
def getOSMTokenForDelete(self):
#JSON_CONTENT_HEADER = {'Content-Type':'application/json'}
JSON_CONTENT_HEADER = {'Accept':'application/json'}
my_type = self.getDBType()
if my_type == 'osm':
print('this SP is a OSM')
sp_host_0 = self.getDBHost()
print (sp_host_0)
sp_host = sp_host_0.__str__()
print (sp_host)
#print (self.getDBHost())
sp_host_1 = sp_host[4:]
print ("sp1 es: ")
print (sp_host_1)
sp_host_2 = sp_host_1[:-10]
print ("sp2 es: ")
print (sp_host_2)
#url = sp_host_2 + '/requests'
url = sp_host_2 + ':9999/osm/admin/v1/tokens'
url_2 = url.replace("http","https")
print (url_2)
#data = request.get_json()
print(url_2)
#print (data)
#print (data['nsd_name'])
#print (data['username'])
#print (data['password'])
#print (data['project_id'])
pr_name = self.getDBProjectName()
print ("project name from DB:")
print (pr_name)
if pr_name:
project_id_for_token = pr_name
if not pr_name:
#data = request.get_json()
#project_id_for_token = data['project_id']
project_id_for_token = self.getDBProject(self)
print ("project name from json body:")
print (pr_name)
#project_id_for_token = data['project_id']
print (project_id_for_token)
#username_for_token = data['username']
#password_for_token = data['password']
username_for_token = self.getDBUserName()
password_for_token = self.getDBPassword()
#project_id_for_token = data['project_id']
admin_data = "{username: 'admin', password: 'admin', project_id: 'admin'}"
print (admin_data)
#update_token = "UPDATE service_platforms SET service_token = \'" +token+ "\' WHERE name = \'" +self.name+ "\'"
data_for_token= "{username: \'" +username_for_token+ "\', password: \'" +password_for_token+ "\', project_id: \'" +project_id_for_token+ "\'}"
#print (data)
get_token = requests.post(url_2,data=data_for_token,headers=JSON_CONTENT_HEADER,verify=False)
print (get_token.text)
print (get_token.content)
token_id = get_token.json()
print (token_id['id'])
#upd_tok = self.updateToken(token_id['id'])
#print (upd_tok)
return token_id['id'] | 34.288916 | 174 | 0.488333 | 6,216 | 56,611 | 4.148327 | 0.041828 | 0.108664 | 0.063988 | 0.04134 | 0.877608 | 0.856395 | 0.842938 | 0.836539 | 0.824556 | 0.8187 | 0 | 0.029934 | 0.402802 | 56,611 | 1,651 | 175 | 34.288916 | 0.732785 | 0.08622 | 0 | 0.842472 | 0 | 0 | 0.095298 | 0.014614 | 0 | 0 | 0 | 0 | 0 | 1 | 0.033943 | false | 0.017406 | 0.008703 | 0.002611 | 0.088773 | 0.382942 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
98a299fc11240dce08dded4ed8e707340fd1e252 | 21,372 | py | Python | tests/all_tests.py | tjweisman/dixit | 57933a78200a8738b973a021ec8d3c3b832d1f7a | [
"MIT"
] | null | null | null | tests/all_tests.py | tjweisman/dixit | 57933a78200a8738b973a021ec8d3c3b832d1f7a | [
"MIT"
] | null | null | null | tests/all_tests.py | tjweisman/dixit | 57933a78200a8738b973a021ec8d3c3b832d1f7a | [
"MIT"
] | null | null | null | from selenium import webdriver
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.common.by import By
import pages
import subtests
def login():
with webdriver.Firefox() as driver:
driver.get(pages.DIXIT_URL)
start_page = pages.StartPage(driver)
start_page.delete_users()
pregame = start_page.join_game("test_user", "test_room")
return "test_user" in pregame.users_list()
def login_multiple():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2"]
session = pages.create_game_session(driver, users, "test_room")
pregame = pages.PregamePage(session.driver)
result_users = pregame.users_list()
return all([user in result_users for user in users])
def two_player_single_round():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
game_correct = subtests.play_all_correct_round(session, gameplay,
[0, 1], users, 0, ["0", "2"])
return game_correct
def two_player_double_round():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
round1_correct = subtests.play_all_correct_round(session, gameplay,
[0,1], users, 0, ["0", "2"])
round2_correct = subtests.play_all_correct_round(session, gameplay,
[0,1], users, 1, ["2", "2"])
return round1_correct and round2_correct
def three_player_single_round():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
p2_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
session.activate(1)
gameplay.click_cid(p2_card)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '0', '4'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid',
'inactive',
'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_3'], [], ['test_2']])
return scores_correct and cards_correct and choices_correct
def leave_before_prompt():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
session.activate(2)
gameplay.leave_game()
session.activate(0)
gameplay.click_hand_card(0)
gameplay.submit_hint()
session.activate(1)
gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.click_allowable_table_card()
gameplay.submit_guess()
users = ['test_1', 'test_2']
scores_correct = (gameplay.get_scores(users) == ['0', '2'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive'])
choices_correct = (gameplay.get_player_choices(users) == [['test_2'], []])
return scores_correct and cards_correct and choices_correct
def leave_after_prompt():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
session.activate(0)
gameplay.click_hand_card(0)
gameplay.submit_hint()
session.activate(2)
gameplay.leave_game()
session.activate(1)
gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.click_allowable_table_card()
gameplay.submit_guess()
users = ['test_1', 'test_2']
scores_correct = (gameplay.get_scores(users) == ['0', '2'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid', 'inactive'])
choices_correct = (gameplay.get_player_choices(users) == [['test_2'], []])
return scores_correct and cards_correct and choices_correct
def leave_after_secret():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.submit_hint()
session.activate(2)
gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.leave_game()
session.activate(1)
gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
remaining_users = ['test_1', 'test_2']
scores_correct = (gameplay.get_scores(remaining_users) == ['0', '2'])
cards_correct = (gameplay.get_card_states(users) ==
['correct-uid', 'inactive', 'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_2'], [], []])
return scores_correct and cards_correct and choices_correct
def leave_before_guess():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.submit_hint()
session.activate(2)
gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(1)
gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
gameplay.leave_game()
session.activate(1)
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
remaining_users = ['test_1', 'test_2']
scores_correct = (gameplay.get_scores(remaining_users) == ['0', '2'])
cards_correct = (gameplay.get_card_states(users) ==
['correct-uid', 'inactive', 'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_2'], [], []])
return scores_correct and cards_correct and choices_correct
def leave_after_guess():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.submit_hint()
session.activate(2)
p3_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(1)
gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
gameplay.leave_game()
session.activate(1)
gameplay.click_cid(p3_card)
gameplay.submit_guess()
remaining_users = ['test_1', 'test_2']
scores_correct = (gameplay.get_scores(remaining_users) == ['3', '0'])
cards_correct = (gameplay.get_card_states(users) ==
['correct-uid', 'inactive', 'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_3'], [], ['test_2']])
return scores_correct and cards_correct and choices_correct
def refresh_before_prompt():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
startpage = session.login_page()
session.activate(2)
session.refresh()
startpage.join_game("test_3", "test_room")
return subtests.play_all_correct_round(session, gameplay, [0,1,2], users,
0, ["0", "2", "2"])
def refresh_after_prompt():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
startpage = session.login_page()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(2)
session.refresh()
startpage.join_game("test_3", "test_room")
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
p2_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
session.activate(1)
gameplay.click_cid(p2_card)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '0', '4'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid',
'inactive',
'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_3'], [], ['test_2']])
return scores_correct and cards_correct and choices_correct
def refresh_after_secret():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
startpage = session.login_page()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(2)
p2_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.refresh()
startpage.join_game("test_3", "test_room")
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
session.activate(1)
gameplay.click_cid(p2_card)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '0', '4'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid',
'inactive',
'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_3'], [], ['test_2']])
return scores_correct and cards_correct and choices_correct
def refresh_before_guess():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
startpage = session.login_page()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(2)
p2_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
session.refresh()
startpage.join_game("test_3", "test_room")
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
session.activate(1)
gameplay.click_cid(p2_card)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '0', '4'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid',
'inactive',
'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_3'], [], ['test_2']])
return scores_correct and cards_correct and choices_correct
def refresh_after_guess():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
gameplay = session.start_game()
startpage = session.login_page()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(2)
p2_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(1)
gameplay.click_cid(p2_card)
gameplay.submit_guess()
session.activate(2)
session.refresh()
startpage.join_game("test_3", "test_room")
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '0', '4'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid',
'inactive',
'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_3'], [], ['test_2']])
return scores_correct and cards_correct and choices_correct
def rejoin_before_prompt():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
pregame = session.pregame_page()
session.activate(2)
pregame.leave_game()
session.activate(0)
gameplay = session.start_game()
startpage = session.login_page()
session.activate(2)
startpage.join_game("test_3", "test_room")
return subtests.play_all_correct_round(session, gameplay, [0,1,2], users,
0, ["0", "2", "2"])
def rejoin_after_prompt():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
pregame = session.pregame_page()
session.activate(2)
pregame.leave_game()
session.activate(0)
gameplay = session.start_game()
startpage = session.login_page()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(2)
startpage.join_game("test_3", "test_room")
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
p2_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
session.activate(1)
gameplay.click_cid(p2_card)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '0', '4'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid',
'inactive',
'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_3'], [], ['test_2']])
return scores_correct and cards_correct and choices_correct
def rejoin_after_secret():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3", "test_4"]
session = pages.create_game_session(driver, users, "test_room")
pregame = session.pregame_page()
session.activate(3)
pregame.leave_game()
session.activate(0)
gameplay = session.start_game()
startpage = session.login_page()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(3)
startpage.join_game("test_4", "test_room")
p3_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
p2_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
session.activate(1)
gameplay.click_cid(p3_card)
gameplay.submit_guess()
session.activate(3)
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '0', '3', '4'])
cards_correct = (gameplay.get_card_states(users) == ['correct-uid',
'inactive',
'inactive',
'inactive'])
choices_correct = (gameplay.get_player_choices(users) ==
[['test_3', 'test_4'], [], [], ['test_2']])
return scores_correct and cards_correct and choices_correct
def rejoin_before_guess():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3"]
session = pages.create_game_session(driver, users, "test_room")
pregame = session.pregame_page()
session.activate(2)
pregame.leave_game()
session.activate(0)
gameplay = session.start_game()
startpage = session.login_page()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
startpage.join_game("test_3", "test_room")
gameplay.click_cid(p1_card)
gameplay.submit_guess()
session.activate(1)
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '4', '0'])
secret_users = ["test_1", "test_2"]
cards_correct = (gameplay.get_card_states(secret_users) == ['correct-uid',
'inactive'])
choices_correct = (gameplay.get_player_choices(secret_users) ==
[['test_2'], ['test_3']])
return scores_correct and cards_correct and choices_correct
def rejoin_after_guess():
with webdriver.Firefox() as driver:
users = ["test_1", "test_2", "test_3", "test_4"]
session = pages.create_game_session(driver, users, "test_room")
pregame = session.pregame_page()
session.activate(3)
pregame.leave_game()
session.activate(0)
gameplay = session.start_game()
startpage = session.login_page()
session.activate(0)
correct_cid = gameplay.click_hand_card(0)
gameplay.set_hint_text("test hint")
gameplay.submit_hint()
session.activate(1)
p1_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
session.activate(2)
p2_card = gameplay.click_hand_card(0)
gameplay.submit_secret()
gameplay.click_cid(p1_card)
gameplay.submit_guess()
session.activate(3)
startpage.join_game("test_4", "test_room")
gameplay.click_cid(correct_cid)
gameplay.submit_guess()
session.activate(1)
gameplay.click_cid(p2_card)
gameplay.submit_guess()
scores_correct = (gameplay.get_scores(users) == ['3', '1', '1', '3'])
secret_users = ["test_1", "test_2", "test_3"]
cards_correct = (gameplay.get_card_states(secret_users) == ['correct-uid',
'inactive',
'inactive'])
choices_correct = (gameplay.get_player_choices(secret_users) ==
[['test_4'], ['test_3'], ['test_2']])
return scores_correct and cards_correct and choices_correct
| 33.709779 | 88 | 0.588106 | 2,351 | 21,372 | 5.041684 | 0.037006 | 0.094913 | 0.063781 | 0.070868 | 0.955032 | 0.95267 | 0.950477 | 0.947355 | 0.946174 | 0.94474 | 0 | 0.021561 | 0.296884 | 21,372 | 633 | 89 | 33.763033 | 0.767219 | 0 | 0 | 0.89852 | 0 | 0 | 0.066629 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.042283 | false | 0 | 0.010571 | 0 | 0.095137 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
7f2d580dba7c737f6ab1c63302ec61219ac2dd47 | 8,965 | py | Python | luiz_viana/graficos_luiz_viana.py | lfelipeviana/visualizacao_covid | 9e061e2260c585e84956225d090838fdb3ee9e82 | [
"MIT"
] | null | null | null | luiz_viana/graficos_luiz_viana.py | lfelipeviana/visualizacao_covid | 9e061e2260c585e84956225d090838fdb3ee9e82 | [
"MIT"
] | null | null | null | luiz_viana/graficos_luiz_viana.py | lfelipeviana/visualizacao_covid | 9e061e2260c585e84956225d090838fdb3ee9e82 | [
"MIT"
] | null | null | null | import dash_core_components as dcc
import dash_html_components as html
from dash.dependencies import Input, Output
import pandas as pd
from .barras_fx_etaria import barras_fx_etaria, barras_sub_plot_vacina
def grafico_total_vacinas_idade(app):
df = pd.read_csv("luiz_viana/data_set_consolidado_vacina.csv")
df = df[df['paciente_endereco_uf'] != 'XX']
estados = df['paciente_endereco_uf'].unique().tolist()
estados.sort()
estados = ['Todos'] + estados
opcoes = [{'label': i, 'value': i} for i in estados]
sexo = ['Masculino', 'Feminino']
opcoes_sexo = [{'label': i, 'value': i[0]} for i in sexo]
fx_etaria = df['faixa_etaria'].unique().tolist()
fx_etaria = [int(item) for item in fx_etaria]
fx_etaria.sort()
opcoes_fx_etaria = [{'label': 'de {} a {}'.format(i, i+10 ), 'value': i}
for i in fx_etaria]
component_html = html.Div([
html.Hr(),
html.H4('Vacinas aplicadas por dia e faixa etária'),
html.Label("Filtros"),
html.Div(
[
html.Div([
html.Label("Sexo:"),
html.Label("Faixa Etária:"),
#html.Label("Estado"),
], style={'width': '10%', 'display': 'inline-block'}),
html.Div(
[
dcc.Checklist(
id='sexo-filter',
options=opcoes_sexo,
value=['M','F'],
labelStyle={'display': 'inline-block'}
),
dcc.Checklist(
id='fx_etaria-filter',
options=opcoes_fx_etaria,
value=fx_etaria,
labelStyle={'display': 'inline-block'}
),
], style={'width': '90%', 'display': 'inline-block'}),
], className="row"),
html.Div(
[
html.Div(
[
html.Label("Estado"),
], style={'width': '10%', 'display': 'inline-block'}),
html.Div([
dcc.Dropdown(
id='estado-filter',
options=opcoes,
value='Todos',
style={"width": "40%"}
),
], style={'width': '90%', 'display': 'inline-block'}),
]),
html.Div(
[
dcc.Graph(id='vacinas-fx_etaria')
])
])
@app.callback(
Output('vacinas-fx_etaria', 'figure'),
Input('estado-filter', 'value'),
Input('sexo-filter', 'value'),
Input('fx_etaria-filter', 'value'))
def update_dataset_fx_etaria(estado, sexo, fx_etaria):
df = pd.read_csv("luiz_viana/data_set_consolidado_vacina.csv")
df = df[df['paciente_endereco_uf'] != 'XX']
r_min = 0
r_max = df.groupby(['vacina_dataaplicacao'],
as_index=False)['quantidade'].sum()['quantidade'].max() + 100000
fx_etaria_opts = df['faixa_etaria'].unique().tolist()
if estado != "Todos":
df = df[df['paciente_endereco_uf'] == estado]
r_min = 0
r_max = df.groupby(['vacina_dataaplicacao'],
as_index=False)['quantidade'].sum()['quantidade'].max() + 100000
for sx in ('M','F'):
if sx not in sexo:
df = df[df['paciente_enumsexobiologico'] != sx]
if fx_etaria is not None:
for fx in fx_etaria_opts:
if fx not in fx_etaria:
df = df[df['faixa_etaria'] != fx]
data_fx = df.groupby(['vacina_dataaplicacao', 'faixa_etaria'],
as_index=False)['quantidade'].sum()
fx0 = [{'vacina_dataaplicacao': '2021-01-31', 'faixa_etaria': 0,
'quantidade': 0}]
fx100 = [{'vacina_dataaplicacao': '2021-01-31', 'faixa_etaria': 100,
'quantidade': 0}]
data_fx.loc[len(data_fx.index)]=list(fx0[0].values())
data_fx.loc[len(data_fx.index)]=list(fx100[0].values())
fig = barras_fx_etaria(data_fx, r_min, r_max)
return fig
return component_html
def grafico_sub_plot_vacinas_idade(app):
df = pd.read_csv("luiz_viana/data_set_consolidado_vacina.csv")
df = df[df['paciente_endereco_uf'] != 'XX']
estados = df['paciente_endereco_uf'].unique().tolist()
estados.sort()
estados = ['Todos'] + estados
opcoes = [{'label': i, 'value': i} for i in estados]
sexo = ['Masculino', 'Feminino']
opcoes_sexo = [{'label': i, 'value': i[0]} for i in sexo]
fx_etaria = df['faixa_etaria'].unique().tolist()
fx_etaria = [int(item) for item in fx_etaria]
fx_etaria.sort()
opcoes_fx_etaria = [{'label': 'de {} a {}'.format(i, i+10 ), 'value': i}
for i in fx_etaria]
component_html = html.Div([
html.Hr(),
html.H4('Vacinas aplicadas por dia e faixa etária'),
html.Label("Filtros"),
html.Div(
[
html.Div([
html.Label("Sexo:"),
html.Label("Faixa Etária:"),
#html.Label("Estado"),
], style={'width': '10%', 'display': 'inline-block'}),
html.Div(
[
dcc.Checklist(
id='sexo-filter',
options=opcoes_sexo,
value=['M','F'],
labelStyle={'display': 'inline-block'}
),
dcc.Checklist(
id='fx_etaria-filter',
options=opcoes_fx_etaria,
value=fx_etaria,
labelStyle={'display': 'inline-block'}
),
], style={'width': '90%', 'display': 'inline-block'}),
], className="row"),
html.Div(
[
html.Div(
[
html.Label("Estado"),
], style={'width': '10%', 'display': 'inline-block'}),
html.Div([
dcc.Dropdown(
id='estado-filter',
options=opcoes,
value='Todos',
style={"width": "40%"}
),
], style={'width': '90%', 'display': 'inline-block'}),
]),
html.Div(
[
dcc.Graph(id='vacinas-fx_etaria')
])
])
@app.callback(
Output('vacinas-fx_etaria', 'figure'),
Input('estado-filter', 'value'),
Input('sexo-filter', 'value'),
Input('fx_etaria-filter', 'value'))
def update_dataset_fx_etaria(estado, sexo, fx_etaria):
df = pd.read_csv("luiz_viana/data_set_consolidado_vacina.csv")
df = df[df['paciente_endereco_uf'] != 'XX']
r_min = 0
r_max = df.groupby(['vacina_dataaplicacao'],
as_index=False)['quantidade'].sum()['quantidade'].max() * 1.1
fx_etaria_opts = df['faixa_etaria'].unique().tolist()
if estado != "Todos":
df = df[df['paciente_endereco_uf'] == estado]
r_min = 0
r_max = df.groupby(['vacina_dataaplicacao'],
as_index=False)['quantidade'].sum()['quantidade'].max() * 1.1
for sx in ('M','F'):
if sx not in sexo:
df = df[df['paciente_enumsexobiologico'] != sx]
if fx_etaria is not None:
for fx in fx_etaria_opts:
if fx not in fx_etaria:
df = df[df['faixa_etaria'] != fx]
data_fx = df.groupby(['vacina_dataaplicacao', 'faixa_etaria',
'vacina_nome'],
as_index=False)['quantidade'].sum()
fig = barras_sub_plot_vacina(data_fx, r_min, r_max)
return fig
return component_html
def graficos_luiz_viana(app):
components_html = []
#components_html.append(grafico_total_vacinas_idade(app))
components_html.append(grafico_sub_plot_vacinas_idade(app))
return components_html | 34.348659 | 78 | 0.452426 | 877 | 8,965 | 4.425314 | 0.142531 | 0.080392 | 0.055656 | 0.028859 | 0.904406 | 0.881474 | 0.870394 | 0.850812 | 0.836898 | 0.836898 | 0 | 0.015198 | 0.412828 | 8,965 | 261 | 79 | 34.348659 | 0.722074 | 0.010931 | 0 | 0.845361 | 0 | 0 | 0.191427 | 0.024817 | 0 | 0 | 0 | 0 | 0 | 1 | 0.025773 | false | 0 | 0.025773 | 0 | 0.07732 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
7f488b8e381d84ced95304a49f4c70dda2aeb848 | 8,165 | py | Python | PostDiffMixture/simulations_folder/simulation_analysis_scripts/phi_functions.py | SIGKDDanon/SIGKDD2021DeAnonV2 | 76f0373ec42ab55feefed3f4ce4bf4d532b51dd2 | [
"Apache-2.0"
] | null | null | null | PostDiffMixture/simulations_folder/simulation_analysis_scripts/phi_functions.py | SIGKDDanon/SIGKDD2021DeAnonV2 | 76f0373ec42ab55feefed3f4ce4bf4d532b51dd2 | [
"Apache-2.0"
] | null | null | null | PostDiffMixture/simulations_folder/simulation_analysis_scripts/phi_functions.py | SIGKDDanon/SIGKDD2021DeAnonV2 | 76f0373ec42ab55feefed3f4ce4bf4d532b51dd2 | [
"Apache-2.0"
] | null | null | null | import ipdb as pb
import pandas as pd
import numpy as np
from scipy.stats import beta
CUML_KEY = "prop_exploring_ppd_cuml"
SNAPSHOT_KKEY = "exploring_ppd_at_this_n"
def plot_phi(df, num_sims, n, c, ax, es = 0):
"""
get prop in cond 1 for when exploring
"""
# pb.set_trace()
step_sizes = [int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))]
extra_steps = [3,4,5]
phi_init = np.sum(np.abs(beta.rvs(1, 1, size=5000) - beta.rvs(1, 1, size=5000)) < c)/5000
phi_list = [phi_init] #per step sizes
for num_steps in step_sizes:#loop over step sizes, make a hist for each
print("num_steps", num_steps)
# df[df["num_steps"] == 16]["exploring_ppd_at_this_n"].mean()
df_for_num_steps = df[df['num_steps'] == num_steps].dropna()
phi = df_for_num_steps[SNAPSHOT_KKEY].mean()
phi_list.append(phi)
# pb.set_trace()
phi_list = np.array(phi_list)
se = np.sqrt(phi_list*(1-phi_list))/np.sqrt(num_sims)
# h = se * stats.t.ppf((1 + 0.95) / 2., num_sims-1)
# h = stats.t.ppf(1-0.025, num_sims)*np.sqrt(phi_list*(1-phi_list)/num_sims) #95 CI for Proportion
#print(phi_list, num_sims)
step_sizes = [0, int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))]
if c == 0.025: #0.05
phi_list_0 = [0, 0, 0, 0, 0]
phi_list_0 = np.array(phi_list_0)
ax.errorbar(step_sizes, phi_list_0,yerr = None,fmt = ".-", label = "c = 0")
ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c))
# else:
# ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c))
ind = np.arange(len(step_sizes))
ind = step_sizes
ax.set_xticks(ind)
print(step_sizes)
ax.set_xticklabels(step_sizes)
ax.tick_params(axis='x', rotation=45)
eps_dec = step_sizes
eps_dec[0] = eps_dec[0] + 1
print("EPS DEC")
print(1.0/np.array(eps_dec))
eps_dec = np.arange(step_sizes[-1])
ax.errorbar(eps_dec, 1.0/np.sqrt(np.array(eps_dec)), color='black', linestyle='--', linewidth = 0.5)
#ax.tick_params(axis='x')
# ax.set_xlabel("number of participants = \n n/2, n, 2*n, 4*n")
#if es == 0.1:
ax.set_xlabel("number of participants")
ax.set_ylim(-0.05, 1.05)
if es == 0:
ax.set_ylabel("$\hat \phi_t$")
if es != 0:
handles, labels = ax.get_legend_handles_labels()
#ax.legend(handles[::-1], labels[::-1], loc='upper left')
ax.legend(handles[::-1], labels[::-1], bbox_to_anchor=(1.1, 1.03))
# ax.legend(handles[::-1], labels[::-1])
#leg2 = ax.legend(loc = 2)
def plot_phi_www(df, num_sims, n, c, ax = None, es = 0):
"""
get prop in cond 1 for when exploring
"""
# pb.set_trace()
step_sizes = [int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))]
#step_sizes = [int(np.ceil(n/2)), int(n)]
extra_steps = [3,4,5]
phi_init = np.sum(np.abs(beta.rvs(1, 1, size=5000) - beta.rvs(1, 1, size=5000)) < c)/5000
phi_list = [phi_init] #per step sizes
# if c != 0:
for num_steps in step_sizes:#loop over step sizes, make a hist for each
print("num_steps", num_steps)
# df[df["num_steps"] == 16]["exploring_ppd_at_this_n"].mean()
df_for_num_steps = df[df['num_steps'] == num_steps].dropna()
phi = df_for_num_steps[SNAPSHOT_KKEY].mean()
phi_list.append(phi)
# pb.set_trace()
phi_list = np.array(phi_list)
se = np.sqrt(phi_list*(1-phi_list))/np.sqrt(num_sims)
# h = se * stats.t.ppf((1 + 0.95) / 2., num_sims-1)
# h = stats.t.ppf(1-0.025, num_sims)*np.sqrt(phi_list*(1-phi_list)/num_sims) #95 CI for Proportion
#print(phi_list, num_sims)
step_sizes = [0, int(np.ceil(n/2)), int(n)]
step_sizes = [0, int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))]
ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c))
# else:
# ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c))
ind = np.arange(len(step_sizes))
ind = step_sizes
ax.set_xticks(ind)
print(step_sizes)
ax.set_xticklabels(step_sizes)
#ax.tick_params(axis='x', rotation=45)
ax.tick_params(axis='x')
# ax.set_xlabel("number of participants = \n n/2, n, 2*n, 4*n")
ax.set_xlabel("number of participants")
ax.set_ylabel("$\hat \phi$")
ax.set_ylim(-0.05, 1.05)
print('HERHEHREHRHERHEHR')
eps_dec = step_sizes
eps_dec[0] = eps_dec[0] + 1
ax.errorbar(step_sizes, 1.0/np.array(eps_dec), color='black', linestyle='--')
if c == 0.1 or 1:
print("PHI 0")
phi_list_0 = [0, 0, 0, 0, 0]
phi_list_0 = np.array(phi_list_0)
ax.errorbar(step_sizes, phi_list_0,yerr = None,fmt = ".-", label = "c = 0")
if n == 785:
handles, labels = ax.get_legend_handles_labels()
#ax.legend(handles[::-1], labels[::-1], loc='upper left')
ax.legend(handles[::-1], labels[::-1], bbox_to_anchor=(1.1, 1.05))
#leg2 = ax.legend(loc = 2)
# fig, ax = plt.subplots(1,4, figsize = (12,5))#for phi
# fig.tight_layout(rect=[0, 0.03, 1, 0.87])
# save_dir = "../simulation_analysis_saves/phi_plots_WWW/NoEffect/"
# Path(save_dir).mkdir(parents=True, exist_ok=True)
#
# title = "$\hat \phi$ Across {} Simulations $p_1 = p_2 = 0.5$ \n $\phi$ := p($|p_1 - p_2| < c$) \n = {}".format(num_sims, n)
#
# fig.suptitle(title)
# fig.savefig(save_dir + "/" + title +".png")
def plot_phi_www_multi(df, df_ne, num_sims, n, c, ax = None, es = 0):
"""
get prop in cond 1 for when exploring
"""
# pb.set_trace()
step_sizes = [int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))]
#step_sizes = [int(np.ceil(n/2)), int(n)]
extra_steps = [3,4,5]
phi_init = np.sum(np.abs(beta.rvs(1, 1, size=5000) - beta.rvs(1, 1, size=5000)) < c)/5000
phi_list = [phi_init] #per step sizes
for num_steps in step_sizes:#loop over step sizes, make a hist for each
print("num_steps", num_steps)
# df[df["num_steps"] == 16]["exploring_ppd_at_this_n"].mean()
df_for_num_steps = df[df['num_steps'] == num_steps].dropna()
phi = df_for_num_steps[SNAPSHOT_KKEY].mean()
phi_list.append(phi)
# pb.set_trace()
phi_list = np.array(phi_list)
se = np.sqrt(phi_list*(1-phi_list))/np.sqrt(num_sims)
# h = se * stats.t.ppf((1 + 0.95) / 2., num_sims-1)
# h = stats.t.ppf(1-0.025, num_sims)*np.sqrt(phi_list*(1-phi_list)/num_sims) #95 CI for Proportion
#print(phi_list, num_sims)
step_sizes = [0, int(np.ceil(n/2)), int(n)]
step_sizes = [0, int(np.ceil(n/2)), int(n), int(np.ceil(2*n)), int(np.ceil(4*n))]
ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c))
# else:
# ax.errorbar(step_sizes, phi_list,yerr = None,fmt = ".-", label = "c = {}".format(c))
ind = np.arange(len(step_sizes))
ind = step_sizes
ax.set_xticks(ind)
print(step_sizes)
ax.set_xticklabels(step_sizes)
#ax.tick_params(axis='x', rotation=45)
ax.tick_params(axis='x')
# ax.set_xlabel("number of participants = \n n/2, n, 2*n, 4*n")
ax.set_xlabel("number of participants")
ax.set_ylabel("$\hat \phi$")
#ax.set_ylim(0.0, 1.0)
ax.set_ylim(-0.05, 1.05)
if n == 785:
handles, labels = ax.get_legend_handles_labels()
#ax.legend(handles[::-1], labels[::-1], loc='upper left')
ax.legend(handles[::-1], labels[::-1], bbox_to_anchor=(1.1, 1.05))
#leg2 = ax.legend(loc = 2)
# fig, ax = plt.subplots(1,4, figsize = (12,5))#for phi
# fig.tight_layout(rect=[0, 0.03, 1, 0.87])
# save_dir = "../simulation_analysis_saves/phi_plots_WWW/NoEffect/"
# Path(save_dir).mkdir(parents=True, exist_ok=True)
#
# title = "$\hat \phi$ Across {} Simulations $p_1 = p_2 = 0.5$ \n $\phi$ := p($|p_1 - p_2| < c$) \n = {}".format(num_sims, n)
#
# fig.suptitle(title)
# fig.savefig(save_dir + "/" + title +".png")
| 39.066986 | 132 | 0.589957 | 1,396 | 8,165 | 3.260029 | 0.113181 | 0.085036 | 0.043507 | 0.026368 | 0.929246 | 0.91782 | 0.905735 | 0.890574 | 0.87673 | 0.87673 | 0 | 0.048283 | 0.218739 | 8,165 | 208 | 133 | 39.254808 | 0.665151 | 0.367789 | 0 | 0.783019 | 0 | 0 | 0.056492 | 0.009118 | 0 | 0 | 0 | 0 | 0 | 1 | 0.028302 | false | 0 | 0.037736 | 0 | 0.066038 | 0.09434 | 0 | 0 | 0 | null | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 7 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.